2019-05-29 23:57:59 +00:00
|
|
|
// SPDX-License-Identifier: GPL-2.0-only
|
2011-03-15 06:53:21 +00:00
|
|
|
/*
|
|
|
|
* Audio and Music Data Transmission Protocol (IEC 61883-6) streams
|
|
|
|
* with Common Isochronous Packet (IEC 61883-1) headers
|
|
|
|
*
|
|
|
|
* Copyright (c) Clemens Ladisch <clemens@ladisch.de>
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <linux/device.h>
|
|
|
|
#include <linux/err.h>
|
|
|
|
#include <linux/firewire.h>
|
2019-10-18 06:19:11 +00:00
|
|
|
#include <linux/firewire-constants.h>
|
2011-03-15 06:53:21 +00:00
|
|
|
#include <linux/module.h>
|
|
|
|
#include <linux/slab.h>
|
|
|
|
#include <sound/pcm.h>
|
2014-04-25 13:44:52 +00:00
|
|
|
#include <sound/pcm_params.h>
|
2015-09-19 02:21:54 +00:00
|
|
|
#include "amdtp-stream.h"
|
2011-03-15 06:53:21 +00:00
|
|
|
|
|
|
|
#define TICKS_PER_CYCLE 3072
|
|
|
|
#define CYCLES_PER_SECOND 8000
|
|
|
|
#define TICKS_PER_SECOND (TICKS_PER_CYCLE * CYCLES_PER_SECOND)
|
|
|
|
|
2021-05-18 13:00:43 +00:00
|
|
|
#define OHCI_SECOND_MODULUS 8
|
2020-05-08 04:36:27 +00:00
|
|
|
|
ALSA: firewire-lib: add tracepoints to dump a part of isochronous packet data
When audio and music units have some quirks in their sequence of packet,
it's really hard for non-owners to identify the quirks. Although developers
need dumps for sequence of packets, it's difficult for users who have no
knowledges and no equipments for this purpose.
This commit adds tracepoints for this situation. When users encounter
the issue, they can dump a part of packet data via Linux tracing framework
as long as using drivers in ALSA firewire stack.
Additionally, tracepoints for outgoing packets will be our help to check
and debug packet processing of ALSA firewire stack.
This commit newly adds 'snd_firewire_lib' subsystem with 'in_packet' and
'out_packet' events. In the events, some attributes of packets and the
index of packet managed by this module are recorded per packet.
This is an usage:
$ trace-cmd record -e snd_firewire_lib:out_packet \
-e snd_firewire_lib:in_packet
/sys/kernel/tracing/events/snd_firewire_lib/out_packet/filter
/sys/kernel/tracing/events/snd_firewire_lib/in_packet/filter
Hit Ctrl^C to stop recording
^C
$ trace-cmd report trace.dat
...
23647.033934: in_packet: 01 4073 ffc0 ffc1 00 000f0040 9001b2d1 122 44
23647.033936: in_packet: 01 4074 ffc0 ffc1 00 000f0048 9001c83b 122 45
23647.033937: in_packet: 01 4075 ffc0 ffc1 00 000f0050 9001ffff 002 46
23647.033938: in_packet: 01 4076 ffc0 ffc1 00 000f0050 9001e1a6 122 47
23647.035426: out_packet: 01 4123 ffc1 ffc0 01 010f00d0 9001fb40 122 17
23647.035428: out_packet: 01 4124 ffc1 ffc0 01 010f00d8 9001ffff 002 18
23647.035429: out_packet: 01 4125 ffc1 ffc0 01 010f00d8 900114aa 122 19
23647.035430: out_packet: 01 4126 ffc1 ffc0 01 010f00e0 90012a15 122 20
(Here, some common fields are omitted so that a line to be within 80
characters.)
...
One line represent one packet. The legend for the last nine fields is:
- The second of cycle scheduled for the packet
- The count of cycle scheduled for the packet
- The ID of node as source (hex)
- Some devices transfer packets with invalid source node ID in their CIP
header.
- The ID of node as destination (hex)
- The value is not in CIP header of packets.
- The value of isochronous channel
- The first quadlet of CIP header (hex)
- The second quadlet of CIP header (hex)
- The number of included quadlets
- The index of packet in a buffer maintained by this module
This is an example to parse these lines from text file by Python3 script:
\#!/usr/bin/env python3
import sys
def parse_ts(second, cycle, syt):
offset = syt & 0xfff
syt >>= 12
if cycle & 0x0f > syt:
cycle += 0x10
cycle &= 0x1ff0
cycle |= syt
second += cycle // 8000
cycle %= 8000
# In CYCLE_TIMER of 1394 OHCI, second is represented in 8 bit.
second %= 128
return (second, cycle, offset)
def calc_ts(second, cycle, offset):
ts = offset
ts += cycle * 3072
# In DMA descriptor of 1394 OHCI, second is represented in 3 bit.
ts += (second % 8) * 8000 * 3072
return ts
def subtract_ts(minuend, subtrahend):
# In DMA descriptor of 1394 OHCI, second is represented in 3 bit.
if minuend < subtrahend:
minuend += 8 * 8000 * 3072
return minuend - subtrahend
if len(sys.argv) != 2:
print('At least, one argument is required for packet dump.')
sys.exit()
filename = sys.argv[1]
data = []
prev = 0
with open(filename, 'r') as f:
for line in f:
pos = line.find('packet:')
if pos < 0:
continue
pos += len('packet:')
line = line[pos:].strip()
fields = line.split(' ')
datum = []
datum.append(fields[8])
syt = int(fields[6][4:], 16)
# Empty packet in IEC 61883-1, or NODATA in IEC 61883-6
if syt == 0xffff:
data_blocks = 0
else:
payload_size = int(fields[7], 10)
data_block_size = int(fields[5][2:4], 16)
data_blocks = (payload_size - 2) / data_block_size
datum.append(data_blocks)
second = int(fields[0], 10)
cycle = int(fields[1], 10)
start = (second << 25) | (cycle << 12)
datum.append('0x{0:08x}'.format(start))
start = calc_ts(second, cycle, 0)
datum.append("0x" + fields[5])
datum.append("0x" + fields[6])
if syt == 0xffff:
second = 0
cycle = 0
tick = 0
else:
second, cycle, tick = parse_ts(second, cycle, syt)
ts = calc_ts(second, cycle, tick)
datum.append(start)
datum.append(ts)
if ts == 0:
datum.append(0)
datum.append(0)
else:
# Usual case, or a case over 8 seconds.
if ts > start or start > 7 * 8000 * 3072:
datum.append(subtract_ts(ts, start))
if ts > prev or start > 7 * 8000 * 3072:
gap = subtract_ts(ts, prev)
datum.append(gap)
else:
datum.append('backward')
else:
datum.append('invalid')
prev = ts
data.append(datum)
sys.exit()
The data variable includes array with these elements:
- The index of the packet
- The number of data blocks in the packet
- The value of cycle count (hex)
- The value of CIP header 1 (hex)
- The value of CIP header 2 (hex)
- The value of cycle count (tick)
- The value of calculated presentation timestamp (tick)
- The offset between the cycle count and presentation timestamp
- The elapsed ticks from the previous presentation timestamp
Signed-off-by: Takashi Sakamoto <o-takashi@sakamocchi.jp>
Signed-off-by: Takashi Iwai <tiwai@suse.de>
2016-05-09 12:12:46 +00:00
|
|
|
/* Always support Linux tracing subsystem. */
|
|
|
|
#define CREATE_TRACE_POINTS
|
|
|
|
#include "amdtp-stream-trace.h"
|
|
|
|
|
2015-02-21 02:50:17 +00:00
|
|
|
#define TRANSFER_DELAY_TICKS 0x2e00 /* 479.17 microseconds */
|
2011-03-15 06:53:21 +00:00
|
|
|
|
2014-04-25 13:44:43 +00:00
|
|
|
/* isochronous header parameters */
|
|
|
|
#define ISO_DATA_LENGTH_SHIFT 16
|
ALSA: firewire-lib: add no-header packet processing
As long as investigating Fireface 400, IEC 61883-1/6 is not applied to
its packet streaming protocol. Remarks of the specific protocol are:
* Each packet doesn't include CIP headers.
* 64,0 and 128,0 kHz are supported.
* The device doesn't necessarily transmit 8,000 packets per second.
* 0, 1, 2, 3 are used as tag for rx isochronous packets, however 0 is
used for tx isochronous packets.
On the other hand, there's a common feature. The number of data blocks
transferred in a second is the same as sampling transmission frequency.
Current ALSA IEC 61883-1/6 engine already has a method to calculate it and
this driver can utilize it for rx packets, as well as tx packets.
This commit adds support for the transferring protocol. CIP_NO_HEADERS
flag is newly added. When this flag is set:
* Both of 0 (without CIP header) and 1 (with CIP header) are used as tag
to handle incoming isochronous packet.
* 0 (without CIP header) is used as tag to transfer outgoing isochronous
packet.
* Skip CIP header evaluation.
* Use unique way to calculate the quadlets of isochronous packet payload.
In ALSA PCM interface, 128.0 kHz is not supported, and the ALSA
IEC 61883-1/6 engine doesn't support 64.0 kHz. These modes are dropped.
The sequence of rx packet has a remarkable quirk about tag. This will be
described in later commits.
Signed-off-by: Takashi Sakamoto <o-takashi@sakamocchi.jp>
Signed-off-by: Takashi Iwai <tiwai@suse.de>
2017-03-31 13:06:07 +00:00
|
|
|
#define TAG_NO_CIP_HEADER 0
|
2011-03-15 06:53:21 +00:00
|
|
|
#define TAG_CIP 1
|
|
|
|
|
2021-05-20 04:01:47 +00:00
|
|
|
// Common Isochronous Packet (CIP) header parameters. Use two quadlets CIP header when supported.
|
|
|
|
#define CIP_HEADER_QUADLETS 2
|
2015-05-22 14:21:12 +00:00
|
|
|
#define CIP_EOH_SHIFT 31
|
|
|
|
#define CIP_EOH (1u << CIP_EOH_SHIFT)
|
2014-04-25 13:44:43 +00:00
|
|
|
#define CIP_EOH_MASK 0x80000000
|
2015-05-22 14:21:12 +00:00
|
|
|
#define CIP_SID_SHIFT 24
|
|
|
|
#define CIP_SID_MASK 0x3f000000
|
|
|
|
#define CIP_DBS_MASK 0x00ff0000
|
|
|
|
#define CIP_DBS_SHIFT 16
|
2017-03-22 12:30:16 +00:00
|
|
|
#define CIP_SPH_MASK 0x00000400
|
|
|
|
#define CIP_SPH_SHIFT 10
|
2015-05-22 14:21:12 +00:00
|
|
|
#define CIP_DBC_MASK 0x000000ff
|
|
|
|
#define CIP_FMT_SHIFT 24
|
2014-04-25 13:44:43 +00:00
|
|
|
#define CIP_FMT_MASK 0x3f000000
|
2015-05-22 14:21:12 +00:00
|
|
|
#define CIP_FDF_MASK 0x00ff0000
|
|
|
|
#define CIP_FDF_SHIFT 16
|
2021-05-24 03:13:43 +00:00
|
|
|
#define CIP_FDF_NO_DATA 0xff
|
2014-04-25 13:44:43 +00:00
|
|
|
#define CIP_SYT_MASK 0x0000ffff
|
|
|
|
#define CIP_SYT_NO_INFO 0xffff
|
2021-05-27 12:26:09 +00:00
|
|
|
#define CIP_SYT_CYCLE_MODULUS 16
|
2021-05-24 03:13:43 +00:00
|
|
|
#define CIP_NO_DATA ((CIP_FDF_NO_DATA << CIP_FDF_SHIFT) | CIP_SYT_NO_INFO)
|
2014-04-25 13:44:43 +00:00
|
|
|
|
2021-05-20 04:01:47 +00:00
|
|
|
#define CIP_HEADER_SIZE (sizeof(__be32) * CIP_HEADER_QUADLETS)
|
|
|
|
|
2015-09-19 02:21:56 +00:00
|
|
|
/* Audio and Music transfer protocol specific parameters */
|
2015-09-19 02:21:53 +00:00
|
|
|
#define CIP_FMT_AM 0x10
|
2014-04-25 13:44:46 +00:00
|
|
|
#define AMDTP_FDF_NO_DATA 0xff
|
2011-03-15 06:53:21 +00:00
|
|
|
|
2019-05-22 14:17:07 +00:00
|
|
|
// For iso header and tstamp.
|
2021-05-20 04:01:47 +00:00
|
|
|
#define IR_CTX_HEADER_DEFAULT_QUADLETS 2
|
|
|
|
// Add nothing.
|
|
|
|
#define IR_CTX_HEADER_SIZE_NO_CIP (sizeof(__be32) * IR_CTX_HEADER_DEFAULT_QUADLETS)
|
|
|
|
// Add two quadlets CIP header.
|
|
|
|
#define IR_CTX_HEADER_SIZE_CIP (IR_CTX_HEADER_SIZE_NO_CIP + CIP_HEADER_SIZE)
|
2019-03-17 11:25:06 +00:00
|
|
|
#define HEADER_TSTAMP_MASK 0x0000ffff
|
2014-04-25 13:44:45 +00:00
|
|
|
|
2021-05-20 04:01:47 +00:00
|
|
|
#define IT_PKT_HEADER_SIZE_CIP CIP_HEADER_SIZE
|
2019-05-23 15:14:40 +00:00
|
|
|
#define IT_PKT_HEADER_SIZE_NO_CIP 0 // Nothing.
|
|
|
|
|
2021-05-18 08:45:55 +00:00
|
|
|
// The initial firmware of OXFW970 can postpone transmission of packet during finishing
|
|
|
|
// asynchronous transaction. This module accepts 5 cycles to skip as maximum to avoid buffer
|
|
|
|
// overrun. Actual device can skip more, then this module stops the packet streaming.
|
|
|
|
#define IR_JUMBO_PAYLOAD_MAX_SKIP_CYCLES 5
|
|
|
|
|
2024-07-30 19:53:26 +00:00
|
|
|
static void pcm_period_work(struct work_struct *work);
|
|
|
|
|
2011-03-15 06:53:21 +00:00
|
|
|
/**
|
2014-04-25 13:44:42 +00:00
|
|
|
* amdtp_stream_init - initialize an AMDTP stream structure
|
|
|
|
* @s: the AMDTP stream to initialize
|
2011-03-15 06:53:21 +00:00
|
|
|
* @unit: the target of the stream
|
2014-04-25 13:44:44 +00:00
|
|
|
* @dir: the direction of stream
|
2021-05-18 08:45:48 +00:00
|
|
|
* @flags: the details of the streaming protocol consist of cip_flags enumeration-constants.
|
2015-09-19 02:21:55 +00:00
|
|
|
* @fmt: the value of fmt field in CIP header
|
2019-07-22 03:37:09 +00:00
|
|
|
* @process_ctx_payloads: callback handler to process payloads of isoc context
|
2015-09-19 02:22:02 +00:00
|
|
|
* @protocol_size: the size to allocate newly for protocol
|
2011-03-15 06:53:21 +00:00
|
|
|
*/
|
2014-04-25 13:44:42 +00:00
|
|
|
int amdtp_stream_init(struct amdtp_stream *s, struct fw_unit *unit,
|
2021-05-18 08:45:48 +00:00
|
|
|
enum amdtp_stream_direction dir, unsigned int flags,
|
2015-09-19 02:22:02 +00:00
|
|
|
unsigned int fmt,
|
2019-07-22 03:37:09 +00:00
|
|
|
amdtp_stream_process_ctx_payloads_t process_ctx_payloads,
|
2015-09-19 02:22:02 +00:00
|
|
|
unsigned int protocol_size)
|
2011-03-15 06:53:21 +00:00
|
|
|
{
|
2019-07-22 03:37:09 +00:00
|
|
|
if (process_ctx_payloads == NULL)
|
2015-09-19 02:22:02 +00:00
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
s->protocol = kzalloc(protocol_size, GFP_KERNEL);
|
|
|
|
if (!s->protocol)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
2015-02-21 14:54:58 +00:00
|
|
|
s->unit = unit;
|
2014-04-25 13:44:44 +00:00
|
|
|
s->direction = dir;
|
2011-03-15 06:53:21 +00:00
|
|
|
s->flags = flags;
|
|
|
|
s->context = ERR_PTR(-1);
|
|
|
|
mutex_init(&s->mutex);
|
2024-07-30 19:53:26 +00:00
|
|
|
INIT_WORK(&s->period_work, pcm_period_work);
|
2011-03-15 06:57:24 +00:00
|
|
|
s->packet_index = 0;
|
2011-03-15 06:53:21 +00:00
|
|
|
|
2021-05-20 04:01:54 +00:00
|
|
|
init_waitqueue_head(&s->ready_wait);
|
2014-04-25 13:44:49 +00:00
|
|
|
|
2015-09-19 02:21:55 +00:00
|
|
|
s->fmt = fmt;
|
2019-07-22 03:37:09 +00:00
|
|
|
s->process_ctx_payloads = process_ctx_payloads;
|
2015-09-19 02:21:53 +00:00
|
|
|
|
2011-03-15 06:53:21 +00:00
|
|
|
return 0;
|
|
|
|
}
|
2014-04-25 13:44:42 +00:00
|
|
|
EXPORT_SYMBOL(amdtp_stream_init);
|
2011-03-15 06:53:21 +00:00
|
|
|
|
|
|
|
/**
|
2014-04-25 13:44:42 +00:00
|
|
|
* amdtp_stream_destroy - free stream resources
|
|
|
|
* @s: the AMDTP stream to destroy
|
2011-03-15 06:53:21 +00:00
|
|
|
*/
|
2014-04-25 13:44:42 +00:00
|
|
|
void amdtp_stream_destroy(struct amdtp_stream *s)
|
2011-03-15 06:53:21 +00:00
|
|
|
{
|
2016-03-30 23:47:02 +00:00
|
|
|
/* Not initialized. */
|
|
|
|
if (s->protocol == NULL)
|
|
|
|
return;
|
|
|
|
|
2014-04-25 13:44:42 +00:00
|
|
|
WARN_ON(amdtp_stream_running(s));
|
2015-09-19 02:22:02 +00:00
|
|
|
kfree(s->protocol);
|
2011-03-15 06:53:21 +00:00
|
|
|
mutex_destroy(&s->mutex);
|
|
|
|
}
|
2014-04-25 13:44:42 +00:00
|
|
|
EXPORT_SYMBOL(amdtp_stream_destroy);
|
2011-03-15 06:53:21 +00:00
|
|
|
|
2011-10-16 19:39:00 +00:00
|
|
|
const unsigned int amdtp_syt_intervals[CIP_SFC_COUNT] = {
|
2011-09-04 20:16:10 +00:00
|
|
|
[CIP_SFC_32000] = 8,
|
|
|
|
[CIP_SFC_44100] = 8,
|
|
|
|
[CIP_SFC_48000] = 8,
|
|
|
|
[CIP_SFC_88200] = 16,
|
|
|
|
[CIP_SFC_96000] = 16,
|
|
|
|
[CIP_SFC_176400] = 32,
|
|
|
|
[CIP_SFC_192000] = 32,
|
|
|
|
};
|
|
|
|
EXPORT_SYMBOL(amdtp_syt_intervals);
|
|
|
|
|
2014-05-27 15:14:36 +00:00
|
|
|
const unsigned int amdtp_rate_table[CIP_SFC_COUNT] = {
|
2014-04-25 13:44:59 +00:00
|
|
|
[CIP_SFC_32000] = 32000,
|
|
|
|
[CIP_SFC_44100] = 44100,
|
|
|
|
[CIP_SFC_48000] = 48000,
|
|
|
|
[CIP_SFC_88200] = 88200,
|
|
|
|
[CIP_SFC_96000] = 96000,
|
|
|
|
[CIP_SFC_176400] = 176400,
|
|
|
|
[CIP_SFC_192000] = 192000,
|
|
|
|
};
|
|
|
|
EXPORT_SYMBOL(amdtp_rate_table);
|
|
|
|
|
2018-09-30 19:11:49 +00:00
|
|
|
static int apply_constraint_to_size(struct snd_pcm_hw_params *params,
|
|
|
|
struct snd_pcm_hw_rule *rule)
|
|
|
|
{
|
|
|
|
struct snd_interval *s = hw_param_interval(params, rule->var);
|
|
|
|
const struct snd_interval *r =
|
|
|
|
hw_param_interval_c(params, SNDRV_PCM_HW_PARAM_RATE);
|
2018-10-30 06:31:15 +00:00
|
|
|
struct snd_interval t = {0};
|
|
|
|
unsigned int step = 0;
|
2018-09-30 19:11:49 +00:00
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < CIP_SFC_COUNT; ++i) {
|
2018-10-30 06:31:15 +00:00
|
|
|
if (snd_interval_test(r, amdtp_rate_table[i]))
|
|
|
|
step = max(step, amdtp_syt_intervals[i]);
|
2018-09-30 19:11:49 +00:00
|
|
|
}
|
|
|
|
|
2024-10-18 06:00:18 +00:00
|
|
|
if (step == 0)
|
|
|
|
return -EINVAL;
|
|
|
|
|
2018-10-30 06:31:15 +00:00
|
|
|
t.min = roundup(s->min, step);
|
|
|
|
t.max = rounddown(s->max, step);
|
|
|
|
t.integer = 1;
|
2018-09-30 19:11:49 +00:00
|
|
|
|
|
|
|
return snd_interval_refine(s, &t);
|
|
|
|
}
|
|
|
|
|
2014-04-25 13:44:52 +00:00
|
|
|
/**
|
|
|
|
* amdtp_stream_add_pcm_hw_constraints - add hw constraints for PCM substream
|
|
|
|
* @s: the AMDTP stream, which must be initialized.
|
|
|
|
* @runtime: the PCM substream runtime
|
|
|
|
*/
|
|
|
|
int amdtp_stream_add_pcm_hw_constraints(struct amdtp_stream *s,
|
|
|
|
struct snd_pcm_runtime *runtime)
|
|
|
|
{
|
2017-06-08 00:11:03 +00:00
|
|
|
struct snd_pcm_hardware *hw = &runtime->hw;
|
2019-10-17 15:54:23 +00:00
|
|
|
unsigned int ctx_header_size;
|
|
|
|
unsigned int maximum_usec_per_period;
|
2014-04-25 13:44:52 +00:00
|
|
|
int err;
|
|
|
|
|
2021-05-27 12:32:53 +00:00
|
|
|
hw->info = SNDRV_PCM_INFO_BLOCK_TRANSFER |
|
2017-06-08 00:11:03 +00:00
|
|
|
SNDRV_PCM_INFO_INTERLEAVED |
|
|
|
|
SNDRV_PCM_INFO_JOINT_DUPLEX |
|
|
|
|
SNDRV_PCM_INFO_MMAP |
|
2021-05-27 12:32:53 +00:00
|
|
|
SNDRV_PCM_INFO_MMAP_VALID |
|
|
|
|
SNDRV_PCM_INFO_NO_PERIOD_WAKEUP;
|
2017-06-08 00:11:03 +00:00
|
|
|
|
|
|
|
hw->periods_min = 2;
|
|
|
|
hw->periods_max = UINT_MAX;
|
|
|
|
|
|
|
|
/* bytes for a frame */
|
|
|
|
hw->period_bytes_min = 4 * hw->channels_max;
|
|
|
|
|
|
|
|
/* Just to prevent from allocating much pages. */
|
|
|
|
hw->period_bytes_max = hw->period_bytes_min * 2048;
|
|
|
|
hw->buffer_bytes_max = hw->period_bytes_max * hw->periods_min;
|
|
|
|
|
2019-10-17 15:54:23 +00:00
|
|
|
// Linux driver for 1394 OHCI controller voluntarily flushes isoc
|
|
|
|
// context when total size of accumulated context header reaches
|
2020-09-09 16:36:59 +00:00
|
|
|
// PAGE_SIZE. This kicks work for the isoc context and brings
|
2019-10-17 15:54:23 +00:00
|
|
|
// callback in the middle of scheduled interrupts.
|
|
|
|
// Although AMDTP streams in the same domain use the same events per
|
|
|
|
// IRQ, use the largest size of context header between IT/IR contexts.
|
|
|
|
// Here, use the value of context header in IR context is for both
|
|
|
|
// contexts.
|
|
|
|
if (!(s->flags & CIP_NO_HEADER))
|
|
|
|
ctx_header_size = IR_CTX_HEADER_SIZE_CIP;
|
|
|
|
else
|
|
|
|
ctx_header_size = IR_CTX_HEADER_SIZE_NO_CIP;
|
|
|
|
maximum_usec_per_period = USEC_PER_SEC * PAGE_SIZE /
|
|
|
|
CYCLES_PER_SECOND / ctx_header_size;
|
|
|
|
|
2019-10-17 15:54:24 +00:00
|
|
|
// In IEC 61883-6, one isoc packet can transfer events up to the value
|
|
|
|
// of syt interval. This comes from the interval of isoc cycle. As 1394
|
|
|
|
// OHCI controller can generate hardware IRQ per isoc packet, the
|
|
|
|
// interval is 125 usec.
|
|
|
|
// However, there are two ways of transmission in IEC 61883-6; blocking
|
|
|
|
// and non-blocking modes. In blocking mode, the sequence of isoc packet
|
|
|
|
// includes 'empty' or 'NODATA' packets which include no event. In
|
|
|
|
// non-blocking mode, the number of events per packet is variable up to
|
|
|
|
// the syt interval.
|
|
|
|
// Due to the above protocol design, the minimum PCM frames per
|
|
|
|
// interrupt should be double of the value of syt interval, thus it is
|
|
|
|
// 250 usec.
|
2014-04-25 13:44:52 +00:00
|
|
|
err = snd_pcm_hw_constraint_minmax(runtime,
|
|
|
|
SNDRV_PCM_HW_PARAM_PERIOD_TIME,
|
2019-10-17 15:54:24 +00:00
|
|
|
250, maximum_usec_per_period);
|
2014-04-25 13:44:52 +00:00
|
|
|
if (err < 0)
|
|
|
|
goto end;
|
|
|
|
|
|
|
|
/* Non-Blocking stream has no more constraints */
|
|
|
|
if (!(s->flags & CIP_BLOCKING))
|
|
|
|
goto end;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* One AMDTP packet can include some frames. In blocking mode, the
|
|
|
|
* number equals to SYT_INTERVAL. So the number is 8, 16 or 32,
|
|
|
|
* depending on its sampling rate. For accurate period interrupt, it's
|
2015-03-09 21:13:03 +00:00
|
|
|
* preferrable to align period/buffer sizes to current SYT_INTERVAL.
|
2014-04-25 13:44:52 +00:00
|
|
|
*/
|
2018-09-30 19:11:49 +00:00
|
|
|
err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_PERIOD_SIZE,
|
|
|
|
apply_constraint_to_size, NULL,
|
2018-10-30 06:31:15 +00:00
|
|
|
SNDRV_PCM_HW_PARAM_PERIOD_SIZE,
|
2018-09-30 19:11:49 +00:00
|
|
|
SNDRV_PCM_HW_PARAM_RATE, -1);
|
|
|
|
if (err < 0)
|
|
|
|
goto end;
|
|
|
|
err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_BUFFER_SIZE,
|
|
|
|
apply_constraint_to_size, NULL,
|
2018-10-30 06:31:15 +00:00
|
|
|
SNDRV_PCM_HW_PARAM_BUFFER_SIZE,
|
2018-09-30 19:11:49 +00:00
|
|
|
SNDRV_PCM_HW_PARAM_RATE, -1);
|
|
|
|
if (err < 0)
|
|
|
|
goto end;
|
2014-04-25 13:44:52 +00:00
|
|
|
end:
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(amdtp_stream_add_pcm_hw_constraints);
|
|
|
|
|
2011-03-15 06:53:21 +00:00
|
|
|
/**
|
2014-04-25 13:44:42 +00:00
|
|
|
* amdtp_stream_set_parameters - set stream parameters
|
|
|
|
* @s: the AMDTP stream to configure
|
2011-03-15 06:53:21 +00:00
|
|
|
* @rate: the sample rate
|
2015-09-19 02:22:02 +00:00
|
|
|
* @data_block_quadlets: the size of a data block in quadlet unit
|
2023-01-10 13:49:31 +00:00
|
|
|
* @pcm_frame_multiplier: the multiplier to compute the number of PCM frames by the number of AMDTP
|
|
|
|
* events.
|
2011-03-15 06:53:21 +00:00
|
|
|
*
|
2011-09-04 20:16:10 +00:00
|
|
|
* The parameters must be set before the stream is started, and must not be
|
2011-03-15 06:53:21 +00:00
|
|
|
* changed while the stream is running.
|
|
|
|
*/
|
2015-09-19 02:22:02 +00:00
|
|
|
int amdtp_stream_set_parameters(struct amdtp_stream *s, unsigned int rate,
|
2023-01-10 13:49:31 +00:00
|
|
|
unsigned int data_block_quadlets, unsigned int pcm_frame_multiplier)
|
2011-03-15 06:53:21 +00:00
|
|
|
{
|
2015-09-19 02:22:02 +00:00
|
|
|
unsigned int sfc;
|
2011-03-15 06:53:21 +00:00
|
|
|
|
2015-09-19 02:21:49 +00:00
|
|
|
for (sfc = 0; sfc < ARRAY_SIZE(amdtp_rate_table); ++sfc) {
|
2014-04-25 13:44:59 +00:00
|
|
|
if (amdtp_rate_table[sfc] == rate)
|
2015-09-19 02:21:49 +00:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
if (sfc == ARRAY_SIZE(amdtp_rate_table))
|
|
|
|
return -EINVAL;
|
2011-09-04 20:12:48 +00:00
|
|
|
|
|
|
|
s->sfc = sfc;
|
2015-09-19 02:22:02 +00:00
|
|
|
s->data_block_quadlets = data_block_quadlets;
|
2011-09-04 20:16:10 +00:00
|
|
|
s->syt_interval = amdtp_syt_intervals[sfc];
|
2011-09-04 20:12:48 +00:00
|
|
|
|
2019-05-21 14:57:34 +00:00
|
|
|
// default buffering in the device.
|
2021-05-22 01:33:03 +00:00
|
|
|
s->transfer_delay = TRANSFER_DELAY_TICKS - TICKS_PER_CYCLE;
|
|
|
|
|
|
|
|
// additional buffering needed to adjust for no-data packets.
|
|
|
|
if (s->flags & CIP_BLOCKING)
|
|
|
|
s->transfer_delay += TICKS_PER_SECOND * s->syt_interval / rate;
|
2014-04-25 13:44:50 +00:00
|
|
|
|
2023-01-10 13:49:31 +00:00
|
|
|
s->pcm_frame_multiplier = pcm_frame_multiplier;
|
|
|
|
|
2015-09-19 02:21:49 +00:00
|
|
|
return 0;
|
2011-03-15 06:53:21 +00:00
|
|
|
}
|
2014-04-25 13:44:42 +00:00
|
|
|
EXPORT_SYMBOL(amdtp_stream_set_parameters);
|
2011-03-15 06:53:21 +00:00
|
|
|
|
2021-05-20 04:01:48 +00:00
|
|
|
// The CIP header is processed in context header apart from context payload.
|
|
|
|
static int amdtp_stream_get_max_ctx_payload_size(struct amdtp_stream *s)
|
|
|
|
{
|
|
|
|
unsigned int multiplier;
|
|
|
|
|
|
|
|
if (s->flags & CIP_JUMBO_PAYLOAD)
|
|
|
|
multiplier = IR_JUMBO_PAYLOAD_MAX_SKIP_CYCLES;
|
|
|
|
else
|
|
|
|
multiplier = 1;
|
|
|
|
|
|
|
|
return s->syt_interval * s->data_block_quadlets * sizeof(__be32) * multiplier;
|
|
|
|
}
|
|
|
|
|
2011-03-15 06:53:21 +00:00
|
|
|
/**
|
2014-04-25 13:44:42 +00:00
|
|
|
* amdtp_stream_get_max_payload - get the stream's packet size
|
|
|
|
* @s: the AMDTP stream
|
2011-03-15 06:53:21 +00:00
|
|
|
*
|
|
|
|
* This function must not be called before the stream has been configured
|
2014-04-25 13:44:42 +00:00
|
|
|
* with amdtp_stream_set_parameters().
|
2011-03-15 06:53:21 +00:00
|
|
|
*/
|
2014-04-25 13:44:42 +00:00
|
|
|
unsigned int amdtp_stream_get_max_payload(struct amdtp_stream *s)
|
2011-03-15 06:53:21 +00:00
|
|
|
{
|
2021-05-20 04:01:48 +00:00
|
|
|
unsigned int cip_header_size;
|
ALSA: firewire-lib: add buffer-over-run protection at receiving more data blocks than expected
In IEC 61883-6, the number of data blocks in a packet is limited up to
the value of SYT_INTERVAL. Current implementation is compliant to the
limitation, while it can cause buffer-over-run when the value of dbs
field in received packet is illegally large.
This commit adds a validator to detect such illegal packets to prevent
the buffer-over-run. Actually, the buffer is aligned to the size of memory
page, thus this issue hardly causes system errors due to the room to page
alignment, as long as a few packets includes such jumbo payload; i.e.
a packet to several received packets.
Here, Behringer F-Control Audio 202 (based on OXFW 960) has a quirk to
postpone transferring isochronous packet till finish handling any
asynchronous packets. In this case, this model is lazy, transfers no
packets according to several cycle-start packets. After finishing, this
model pushes required data in next isochronous packet. As a result, the
packet include more data blocks than IEC 61883-6 defines.
To continue to support this model, this commit adds a new flag to extend
the length of calculated payload. This flag allows the size of payload
5 times as large as IEC 61883-6 defines. As a result, packets from this
model passed the validator successfully.
Signed-off-by: Takashi Sakamoto <o-takashi@sakamocchi.jp>
Signed-off-by: Takashi Iwai <tiwai@suse.de>
2015-05-22 14:00:50 +00:00
|
|
|
|
ALSA: firewire-lib: add no-header packet processing
As long as investigating Fireface 400, IEC 61883-1/6 is not applied to
its packet streaming protocol. Remarks of the specific protocol are:
* Each packet doesn't include CIP headers.
* 64,0 and 128,0 kHz are supported.
* The device doesn't necessarily transmit 8,000 packets per second.
* 0, 1, 2, 3 are used as tag for rx isochronous packets, however 0 is
used for tx isochronous packets.
On the other hand, there's a common feature. The number of data blocks
transferred in a second is the same as sampling transmission frequency.
Current ALSA IEC 61883-1/6 engine already has a method to calculate it and
this driver can utilize it for rx packets, as well as tx packets.
This commit adds support for the transferring protocol. CIP_NO_HEADERS
flag is newly added. When this flag is set:
* Both of 0 (without CIP header) and 1 (with CIP header) are used as tag
to handle incoming isochronous packet.
* 0 (without CIP header) is used as tag to transfer outgoing isochronous
packet.
* Skip CIP header evaluation.
* Use unique way to calculate the quadlets of isochronous packet payload.
In ALSA PCM interface, 128.0 kHz is not supported, and the ALSA
IEC 61883-1/6 engine doesn't support 64.0 kHz. These modes are dropped.
The sequence of rx packet has a remarkable quirk about tag. This will be
described in later commits.
Signed-off-by: Takashi Sakamoto <o-takashi@sakamocchi.jp>
Signed-off-by: Takashi Iwai <tiwai@suse.de>
2017-03-31 13:06:07 +00:00
|
|
|
if (!(s->flags & CIP_NO_HEADER))
|
2021-05-20 04:01:47 +00:00
|
|
|
cip_header_size = CIP_HEADER_SIZE;
|
2021-05-20 04:01:48 +00:00
|
|
|
else
|
|
|
|
cip_header_size = 0;
|
ALSA: firewire-lib: add buffer-over-run protection at receiving more data blocks than expected
In IEC 61883-6, the number of data blocks in a packet is limited up to
the value of SYT_INTERVAL. Current implementation is compliant to the
limitation, while it can cause buffer-over-run when the value of dbs
field in received packet is illegally large.
This commit adds a validator to detect such illegal packets to prevent
the buffer-over-run. Actually, the buffer is aligned to the size of memory
page, thus this issue hardly causes system errors due to the room to page
alignment, as long as a few packets includes such jumbo payload; i.e.
a packet to several received packets.
Here, Behringer F-Control Audio 202 (based on OXFW 960) has a quirk to
postpone transferring isochronous packet till finish handling any
asynchronous packets. In this case, this model is lazy, transfers no
packets according to several cycle-start packets. After finishing, this
model pushes required data in next isochronous packet. As a result, the
packet include more data blocks than IEC 61883-6 defines.
To continue to support this model, this commit adds a new flag to extend
the length of calculated payload. This flag allows the size of payload
5 times as large as IEC 61883-6 defines. As a result, packets from this
model passed the validator successfully.
Signed-off-by: Takashi Sakamoto <o-takashi@sakamocchi.jp>
Signed-off-by: Takashi Iwai <tiwai@suse.de>
2015-05-22 14:00:50 +00:00
|
|
|
|
2021-05-20 04:01:48 +00:00
|
|
|
return cip_header_size + amdtp_stream_get_max_ctx_payload_size(s);
|
2011-03-15 06:53:21 +00:00
|
|
|
}
|
2014-04-25 13:44:42 +00:00
|
|
|
EXPORT_SYMBOL(amdtp_stream_get_max_payload);
|
2011-03-15 06:53:21 +00:00
|
|
|
|
2012-05-13 20:03:09 +00:00
|
|
|
/**
|
2014-04-25 13:44:42 +00:00
|
|
|
* amdtp_stream_pcm_prepare - prepare PCM device for running
|
|
|
|
* @s: the AMDTP stream
|
2012-05-13 20:03:09 +00:00
|
|
|
*
|
|
|
|
* This function should be called from the PCM device's .prepare callback.
|
|
|
|
*/
|
2014-04-25 13:44:42 +00:00
|
|
|
void amdtp_stream_pcm_prepare(struct amdtp_stream *s)
|
2012-05-13 20:03:09 +00:00
|
|
|
{
|
2024-07-30 19:53:26 +00:00
|
|
|
cancel_work_sync(&s->period_work);
|
2012-05-13 20:03:09 +00:00
|
|
|
s->pcm_buffer_pointer = 0;
|
|
|
|
s->pcm_period_pointer = 0;
|
|
|
|
}
|
2014-04-25 13:44:42 +00:00
|
|
|
EXPORT_SYMBOL(amdtp_stream_pcm_prepare);
|
2012-05-13 20:03:09 +00:00
|
|
|
|
2023-01-10 13:49:33 +00:00
|
|
|
#define prev_packet_desc(s, desc) \
|
|
|
|
list_prev_entry_circular(desc, &s->packet_descs_list, link)
|
|
|
|
|
2021-05-22 01:33:02 +00:00
|
|
|
static void pool_blocking_data_blocks(struct amdtp_stream *s, struct seq_desc *descs,
|
2023-01-07 02:32:11 +00:00
|
|
|
unsigned int size, unsigned int pos, unsigned int count)
|
2011-03-15 06:53:21 +00:00
|
|
|
{
|
2021-05-22 01:33:02 +00:00
|
|
|
const unsigned int syt_interval = s->syt_interval;
|
|
|
|
int i;
|
2011-03-15 06:53:21 +00:00
|
|
|
|
2021-05-22 01:33:02 +00:00
|
|
|
for (i = 0; i < count; ++i) {
|
2023-01-07 02:32:11 +00:00
|
|
|
struct seq_desc *desc = descs + pos;
|
2021-05-22 01:33:02 +00:00
|
|
|
|
|
|
|
if (desc->syt_offset != CIP_SYT_NO_INFO)
|
|
|
|
desc->data_blocks = syt_interval;
|
2015-05-22 14:00:51 +00:00
|
|
|
else
|
2021-05-22 01:33:02 +00:00
|
|
|
desc->data_blocks = 0;
|
|
|
|
|
2023-01-07 02:32:11 +00:00
|
|
|
pos = (pos + 1) % size;
|
2021-05-22 01:33:02 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void pool_ideal_nonblocking_data_blocks(struct amdtp_stream *s, struct seq_desc *descs,
|
2023-01-07 02:32:11 +00:00
|
|
|
unsigned int size, unsigned int pos,
|
2021-05-22 01:33:02 +00:00
|
|
|
unsigned int count)
|
|
|
|
{
|
|
|
|
const enum cip_sfc sfc = s->sfc;
|
|
|
|
unsigned int state = s->ctx_data.rx.data_block_state;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < count; ++i) {
|
2023-01-07 02:32:11 +00:00
|
|
|
struct seq_desc *desc = descs + pos;
|
2021-05-22 01:33:02 +00:00
|
|
|
|
2020-05-08 04:36:32 +00:00
|
|
|
if (!cip_sfc_is_base_44100(sfc)) {
|
2019-05-21 14:57:34 +00:00
|
|
|
// Sample_rate / 8000 is an integer, and precomputed.
|
2021-05-22 01:33:02 +00:00
|
|
|
desc->data_blocks = state;
|
2015-05-22 14:00:51 +00:00
|
|
|
} else {
|
2021-05-22 01:33:02 +00:00
|
|
|
unsigned int phase = state;
|
2011-03-15 06:53:21 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* This calculates the number of data blocks per packet so that
|
|
|
|
* 1) the overall rate is correct and exactly synchronized to
|
|
|
|
* the bus clock, and
|
|
|
|
* 2) packets with a rounded-up number of blocks occur as early
|
|
|
|
* as possible in the sequence (to prevent underruns of the
|
|
|
|
* device's buffer).
|
|
|
|
*/
|
2020-05-08 04:36:32 +00:00
|
|
|
if (sfc == CIP_SFC_44100)
|
2015-05-22 14:00:51 +00:00
|
|
|
/* 6 6 5 6 5 6 5 ... */
|
2021-05-22 01:33:02 +00:00
|
|
|
desc->data_blocks = 5 + ((phase & 1) ^ (phase == 0 || phase >= 40));
|
2015-05-22 14:00:51 +00:00
|
|
|
else
|
|
|
|
/* 12 11 11 11 11 ... or 23 22 22 22 22 ... */
|
2021-05-22 01:33:02 +00:00
|
|
|
desc->data_blocks = 11 * (sfc >> 1) + (phase == 0);
|
2020-05-08 04:36:32 +00:00
|
|
|
if (++phase >= (80 >> (sfc >> 1)))
|
2015-05-22 14:00:51 +00:00
|
|
|
phase = 0;
|
2021-05-22 01:33:02 +00:00
|
|
|
state = phase;
|
2015-05-22 14:00:51 +00:00
|
|
|
}
|
2021-05-22 01:33:02 +00:00
|
|
|
|
2023-01-07 02:32:11 +00:00
|
|
|
pos = (pos + 1) % size;
|
2011-03-15 06:53:21 +00:00
|
|
|
}
|
|
|
|
|
2021-05-22 01:33:02 +00:00
|
|
|
s->ctx_data.rx.data_block_state = state;
|
2011-03-15 06:53:21 +00:00
|
|
|
}
|
|
|
|
|
2020-05-08 04:36:31 +00:00
|
|
|
static unsigned int calculate_syt_offset(unsigned int *last_syt_offset,
|
|
|
|
unsigned int *syt_offset_state, enum cip_sfc sfc)
|
2011-03-15 06:53:21 +00:00
|
|
|
{
|
2020-05-08 04:36:31 +00:00
|
|
|
unsigned int syt_offset;
|
2011-03-15 06:53:21 +00:00
|
|
|
|
2020-05-08 04:36:31 +00:00
|
|
|
if (*last_syt_offset < TICKS_PER_CYCLE) {
|
|
|
|
if (!cip_sfc_is_base_44100(sfc))
|
|
|
|
syt_offset = *last_syt_offset + *syt_offset_state;
|
2011-03-15 06:53:21 +00:00
|
|
|
else {
|
|
|
|
/*
|
|
|
|
* The time, in ticks, of the n'th SYT_INTERVAL sample is:
|
|
|
|
* n * SYT_INTERVAL * 24576000 / sample_rate
|
|
|
|
* Modulo TICKS_PER_CYCLE, the difference between successive
|
|
|
|
* elements is about 1386.23. Rounding the results of this
|
|
|
|
* formula to the SYT precision results in a sequence of
|
|
|
|
* differences that begins with:
|
|
|
|
* 1386 1386 1387 1386 1386 1386 1387 1386 1386 1386 1387 ...
|
|
|
|
* This code generates _exactly_ the same sequence.
|
|
|
|
*/
|
2020-05-08 04:36:31 +00:00
|
|
|
unsigned int phase = *syt_offset_state;
|
|
|
|
unsigned int index = phase % 13;
|
|
|
|
|
|
|
|
syt_offset = *last_syt_offset;
|
2011-03-15 06:53:21 +00:00
|
|
|
syt_offset += 1386 + ((index && !(index & 3)) ||
|
|
|
|
phase == 146);
|
|
|
|
if (++phase >= 147)
|
|
|
|
phase = 0;
|
2020-05-08 04:36:31 +00:00
|
|
|
*syt_offset_state = phase;
|
2011-03-15 06:53:21 +00:00
|
|
|
}
|
|
|
|
} else
|
2020-05-08 04:36:31 +00:00
|
|
|
syt_offset = *last_syt_offset - TICKS_PER_CYCLE;
|
|
|
|
*last_syt_offset = syt_offset;
|
2011-03-15 06:53:21 +00:00
|
|
|
|
2020-05-08 04:36:30 +00:00
|
|
|
if (syt_offset >= TICKS_PER_CYCLE)
|
|
|
|
syt_offset = CIP_SYT_NO_INFO;
|
2011-03-15 06:53:21 +00:00
|
|
|
|
2020-05-08 04:36:30 +00:00
|
|
|
return syt_offset;
|
2011-03-15 06:53:21 +00:00
|
|
|
}
|
|
|
|
|
2021-05-22 01:33:01 +00:00
|
|
|
static void pool_ideal_syt_offsets(struct amdtp_stream *s, struct seq_desc *descs,
|
2023-01-07 02:32:11 +00:00
|
|
|
unsigned int size, unsigned int pos, unsigned int count)
|
2021-05-22 01:33:01 +00:00
|
|
|
{
|
|
|
|
const enum cip_sfc sfc = s->sfc;
|
|
|
|
unsigned int last = s->ctx_data.rx.last_syt_offset;
|
|
|
|
unsigned int state = s->ctx_data.rx.syt_offset_state;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < count; ++i) {
|
2023-01-07 02:32:11 +00:00
|
|
|
struct seq_desc *desc = descs + pos;
|
2021-05-22 01:33:01 +00:00
|
|
|
|
|
|
|
desc->syt_offset = calculate_syt_offset(&last, &state, sfc);
|
|
|
|
|
2023-01-07 02:32:11 +00:00
|
|
|
pos = (pos + 1) % size;
|
2021-05-22 01:33:01 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
s->ctx_data.rx.last_syt_offset = last;
|
|
|
|
s->ctx_data.rx.syt_offset_state = state;
|
|
|
|
}
|
|
|
|
|
2021-05-27 12:26:09 +00:00
|
|
|
static unsigned int compute_syt_offset(unsigned int syt, unsigned int cycle,
|
|
|
|
unsigned int transfer_delay)
|
|
|
|
{
|
|
|
|
unsigned int cycle_lo = (cycle % CYCLES_PER_SECOND) & 0x0f;
|
|
|
|
unsigned int syt_cycle_lo = (syt & 0xf000) >> 12;
|
|
|
|
unsigned int syt_offset;
|
|
|
|
|
|
|
|
// Round up.
|
|
|
|
if (syt_cycle_lo < cycle_lo)
|
|
|
|
syt_cycle_lo += CIP_SYT_CYCLE_MODULUS;
|
|
|
|
syt_cycle_lo -= cycle_lo;
|
|
|
|
|
|
|
|
// Subtract transfer delay so that the synchronization offset is not so large
|
|
|
|
// at transmission.
|
|
|
|
syt_offset = syt_cycle_lo * TICKS_PER_CYCLE + (syt & 0x0fff);
|
|
|
|
if (syt_offset < transfer_delay)
|
|
|
|
syt_offset += CIP_SYT_CYCLE_MODULUS * TICKS_PER_CYCLE;
|
|
|
|
|
|
|
|
return syt_offset - transfer_delay;
|
|
|
|
}
|
|
|
|
|
ALSA: firewire-lib: replay sequence of incoming packets for outgoing packets
ALSA IEC 61883-1/6 packet streaming engine uses pre-computed parameters
ideal for nominal sampling transfer frequency (STF) to transfer packets
to device since it was added 2011. As a result of user experience for a
decade, it is clear that the sequence is not suitable to some actual
devices. It takes the devices to generate noise, and causes any type of
discontinuity in the series of packet transferred from the device. It's
required for the engine to transfer packets according to effective STF.
The effective STF is given by media clock recovered by the sequence of
packet transferred from the target device. In the previous commit, the
sequence is already cached. The media clock recovery can be achieved by
analyzing the sequence.
In technological world, many ideas are proposed for media clock recovery.
However, the small part of them could be actually adopted in our case
since floating point arithmetic is not mostly available in Linux kernel
land.
This commit adopts the simple way from them; sequence replay, which means
that the sequence of parameters from incoming packet is used as is to
transfer outgoing packets. The media clock is not computed internally,
but the sequence of outgoing packet superficially looks to be generated by
the media clock.
The association between source and destination is decided when starting
AMDTP domain. When the target device supports a pair of isochronous packet
streams, the tx stream is source and the rx stream is destination. When it
supports two pair of streams, each of tx stream is associated to
corresponding rx stream in its order. When it supports less number of tx
streams than rx streams, the fist tx stream is selected for all of rx
streams. When it supports more tx streams than rx streams, the first tx
packet is associated to the rx stream.
As I noted in previous commit, the sequence of parameters from incoming
packet is different between devices, time to time. It is worse idea to
replay the sequence of parameters from a device for the sequence of
packet to the other devices even if they are in the same category of
device.
Signed-off-by: Takashi Sakamoto <o-takashi@sakamocchi.jp>
Link: https://lore.kernel.org/r/20210527122611.173711-3-o-takashi@sakamocchi.jp
Signed-off-by: Takashi Iwai <tiwai@suse.de>
2021-05-27 12:26:10 +00:00
|
|
|
// Both of the producer and consumer of the queue runs in the same clock of IEEE 1394 bus.
|
|
|
|
// Additionally, the sequence of tx packets is severely checked against any discontinuity
|
|
|
|
// before filling entries in the queue. The calculation is safe even if it looks fragile by
|
|
|
|
// overrun.
|
|
|
|
static unsigned int calculate_cached_cycle_count(struct amdtp_stream *s, unsigned int head)
|
|
|
|
{
|
|
|
|
const unsigned int cache_size = s->ctx_data.tx.cache.size;
|
2023-01-07 02:32:13 +00:00
|
|
|
unsigned int cycles = s->ctx_data.tx.cache.pos;
|
ALSA: firewire-lib: replay sequence of incoming packets for outgoing packets
ALSA IEC 61883-1/6 packet streaming engine uses pre-computed parameters
ideal for nominal sampling transfer frequency (STF) to transfer packets
to device since it was added 2011. As a result of user experience for a
decade, it is clear that the sequence is not suitable to some actual
devices. It takes the devices to generate noise, and causes any type of
discontinuity in the series of packet transferred from the device. It's
required for the engine to transfer packets according to effective STF.
The effective STF is given by media clock recovered by the sequence of
packet transferred from the target device. In the previous commit, the
sequence is already cached. The media clock recovery can be achieved by
analyzing the sequence.
In technological world, many ideas are proposed for media clock recovery.
However, the small part of them could be actually adopted in our case
since floating point arithmetic is not mostly available in Linux kernel
land.
This commit adopts the simple way from them; sequence replay, which means
that the sequence of parameters from incoming packet is used as is to
transfer outgoing packets. The media clock is not computed internally,
but the sequence of outgoing packet superficially looks to be generated by
the media clock.
The association between source and destination is decided when starting
AMDTP domain. When the target device supports a pair of isochronous packet
streams, the tx stream is source and the rx stream is destination. When it
supports two pair of streams, each of tx stream is associated to
corresponding rx stream in its order. When it supports less number of tx
streams than rx streams, the fist tx stream is selected for all of rx
streams. When it supports more tx streams than rx streams, the first tx
packet is associated to the rx stream.
As I noted in previous commit, the sequence of parameters from incoming
packet is different between devices, time to time. It is worse idea to
replay the sequence of parameters from a device for the sequence of
packet to the other devices even if they are in the same category of
device.
Signed-off-by: Takashi Sakamoto <o-takashi@sakamocchi.jp>
Link: https://lore.kernel.org/r/20210527122611.173711-3-o-takashi@sakamocchi.jp
Signed-off-by: Takashi Iwai <tiwai@suse.de>
2021-05-27 12:26:10 +00:00
|
|
|
|
|
|
|
if (cycles < head)
|
|
|
|
cycles += cache_size;
|
|
|
|
cycles -= head;
|
|
|
|
|
|
|
|
return cycles;
|
|
|
|
}
|
|
|
|
|
2023-01-09 02:17:36 +00:00
|
|
|
static void cache_seq(struct amdtp_stream *s, const struct pkt_desc *src, unsigned int desc_count)
|
2021-05-27 12:26:09 +00:00
|
|
|
{
|
|
|
|
const unsigned int transfer_delay = s->transfer_delay;
|
|
|
|
const unsigned int cache_size = s->ctx_data.tx.cache.size;
|
|
|
|
struct seq_desc *cache = s->ctx_data.tx.cache.descs;
|
2023-01-07 02:32:13 +00:00
|
|
|
unsigned int cache_pos = s->ctx_data.tx.cache.pos;
|
2021-05-27 12:26:09 +00:00
|
|
|
bool aware_syt = !(s->flags & CIP_UNAWARE_SYT);
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < desc_count; ++i) {
|
2023-01-07 02:32:13 +00:00
|
|
|
struct seq_desc *dst = cache + cache_pos;
|
2021-05-27 12:26:09 +00:00
|
|
|
|
|
|
|
if (aware_syt && src->syt != CIP_SYT_NO_INFO)
|
|
|
|
dst->syt_offset = compute_syt_offset(src->syt, src->cycle, transfer_delay);
|
|
|
|
else
|
|
|
|
dst->syt_offset = CIP_SYT_NO_INFO;
|
|
|
|
dst->data_blocks = src->data_blocks;
|
|
|
|
|
2023-01-07 02:32:13 +00:00
|
|
|
cache_pos = (cache_pos + 1) % cache_size;
|
2023-01-09 02:17:36 +00:00
|
|
|
src = amdtp_stream_next_packet_desc(s, src);
|
2021-05-27 12:26:09 +00:00
|
|
|
}
|
|
|
|
|
2023-01-07 02:32:13 +00:00
|
|
|
s->ctx_data.tx.cache.pos = cache_pos;
|
2021-05-27 12:26:09 +00:00
|
|
|
}
|
|
|
|
|
2023-01-07 02:32:11 +00:00
|
|
|
static void pool_ideal_seq_descs(struct amdtp_stream *s, struct seq_desc *descs, unsigned int size,
|
|
|
|
unsigned int pos, unsigned int count)
|
2021-05-22 01:32:59 +00:00
|
|
|
{
|
2023-01-07 02:32:11 +00:00
|
|
|
pool_ideal_syt_offsets(s, descs, size, pos, count);
|
2021-05-22 01:33:01 +00:00
|
|
|
|
2021-05-22 01:33:02 +00:00
|
|
|
if (s->flags & CIP_BLOCKING)
|
2023-01-07 02:32:11 +00:00
|
|
|
pool_blocking_data_blocks(s, descs, size, pos, count);
|
2021-05-22 01:33:02 +00:00
|
|
|
else
|
2023-01-07 02:32:11 +00:00
|
|
|
pool_ideal_nonblocking_data_blocks(s, descs, size, pos, count);
|
2021-05-22 01:32:59 +00:00
|
|
|
}
|
|
|
|
|
2023-01-07 02:32:11 +00:00
|
|
|
static void pool_replayed_seq(struct amdtp_stream *s, struct seq_desc *descs, unsigned int size,
|
|
|
|
unsigned int pos, unsigned int count)
|
ALSA: firewire-lib: replay sequence of incoming packets for outgoing packets
ALSA IEC 61883-1/6 packet streaming engine uses pre-computed parameters
ideal for nominal sampling transfer frequency (STF) to transfer packets
to device since it was added 2011. As a result of user experience for a
decade, it is clear that the sequence is not suitable to some actual
devices. It takes the devices to generate noise, and causes any type of
discontinuity in the series of packet transferred from the device. It's
required for the engine to transfer packets according to effective STF.
The effective STF is given by media clock recovered by the sequence of
packet transferred from the target device. In the previous commit, the
sequence is already cached. The media clock recovery can be achieved by
analyzing the sequence.
In technological world, many ideas are proposed for media clock recovery.
However, the small part of them could be actually adopted in our case
since floating point arithmetic is not mostly available in Linux kernel
land.
This commit adopts the simple way from them; sequence replay, which means
that the sequence of parameters from incoming packet is used as is to
transfer outgoing packets. The media clock is not computed internally,
but the sequence of outgoing packet superficially looks to be generated by
the media clock.
The association between source and destination is decided when starting
AMDTP domain. When the target device supports a pair of isochronous packet
streams, the tx stream is source and the rx stream is destination. When it
supports two pair of streams, each of tx stream is associated to
corresponding rx stream in its order. When it supports less number of tx
streams than rx streams, the fist tx stream is selected for all of rx
streams. When it supports more tx streams than rx streams, the first tx
packet is associated to the rx stream.
As I noted in previous commit, the sequence of parameters from incoming
packet is different between devices, time to time. It is worse idea to
replay the sequence of parameters from a device for the sequence of
packet to the other devices even if they are in the same category of
device.
Signed-off-by: Takashi Sakamoto <o-takashi@sakamocchi.jp>
Link: https://lore.kernel.org/r/20210527122611.173711-3-o-takashi@sakamocchi.jp
Signed-off-by: Takashi Iwai <tiwai@suse.de>
2021-05-27 12:26:10 +00:00
|
|
|
{
|
|
|
|
struct amdtp_stream *target = s->ctx_data.rx.replay_target;
|
|
|
|
const struct seq_desc *cache = target->ctx_data.tx.cache.descs;
|
|
|
|
const unsigned int cache_size = target->ctx_data.tx.cache.size;
|
2023-01-07 02:32:14 +00:00
|
|
|
unsigned int cache_pos = s->ctx_data.rx.cache_pos;
|
ALSA: firewire-lib: replay sequence of incoming packets for outgoing packets
ALSA IEC 61883-1/6 packet streaming engine uses pre-computed parameters
ideal for nominal sampling transfer frequency (STF) to transfer packets
to device since it was added 2011. As a result of user experience for a
decade, it is clear that the sequence is not suitable to some actual
devices. It takes the devices to generate noise, and causes any type of
discontinuity in the series of packet transferred from the device. It's
required for the engine to transfer packets according to effective STF.
The effective STF is given by media clock recovered by the sequence of
packet transferred from the target device. In the previous commit, the
sequence is already cached. The media clock recovery can be achieved by
analyzing the sequence.
In technological world, many ideas are proposed for media clock recovery.
However, the small part of them could be actually adopted in our case
since floating point arithmetic is not mostly available in Linux kernel
land.
This commit adopts the simple way from them; sequence replay, which means
that the sequence of parameters from incoming packet is used as is to
transfer outgoing packets. The media clock is not computed internally,
but the sequence of outgoing packet superficially looks to be generated by
the media clock.
The association between source and destination is decided when starting
AMDTP domain. When the target device supports a pair of isochronous packet
streams, the tx stream is source and the rx stream is destination. When it
supports two pair of streams, each of tx stream is associated to
corresponding rx stream in its order. When it supports less number of tx
streams than rx streams, the fist tx stream is selected for all of rx
streams. When it supports more tx streams than rx streams, the first tx
packet is associated to the rx stream.
As I noted in previous commit, the sequence of parameters from incoming
packet is different between devices, time to time. It is worse idea to
replay the sequence of parameters from a device for the sequence of
packet to the other devices even if they are in the same category of
device.
Signed-off-by: Takashi Sakamoto <o-takashi@sakamocchi.jp>
Link: https://lore.kernel.org/r/20210527122611.173711-3-o-takashi@sakamocchi.jp
Signed-off-by: Takashi Iwai <tiwai@suse.de>
2021-05-27 12:26:10 +00:00
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < count; ++i) {
|
2023-01-07 02:32:14 +00:00
|
|
|
descs[pos] = cache[cache_pos];
|
|
|
|
cache_pos = (cache_pos + 1) % cache_size;
|
2023-01-07 02:32:11 +00:00
|
|
|
pos = (pos + 1) % size;
|
ALSA: firewire-lib: replay sequence of incoming packets for outgoing packets
ALSA IEC 61883-1/6 packet streaming engine uses pre-computed parameters
ideal for nominal sampling transfer frequency (STF) to transfer packets
to device since it was added 2011. As a result of user experience for a
decade, it is clear that the sequence is not suitable to some actual
devices. It takes the devices to generate noise, and causes any type of
discontinuity in the series of packet transferred from the device. It's
required for the engine to transfer packets according to effective STF.
The effective STF is given by media clock recovered by the sequence of
packet transferred from the target device. In the previous commit, the
sequence is already cached. The media clock recovery can be achieved by
analyzing the sequence.
In technological world, many ideas are proposed for media clock recovery.
However, the small part of them could be actually adopted in our case
since floating point arithmetic is not mostly available in Linux kernel
land.
This commit adopts the simple way from them; sequence replay, which means
that the sequence of parameters from incoming packet is used as is to
transfer outgoing packets. The media clock is not computed internally,
but the sequence of outgoing packet superficially looks to be generated by
the media clock.
The association between source and destination is decided when starting
AMDTP domain. When the target device supports a pair of isochronous packet
streams, the tx stream is source and the rx stream is destination. When it
supports two pair of streams, each of tx stream is associated to
corresponding rx stream in its order. When it supports less number of tx
streams than rx streams, the fist tx stream is selected for all of rx
streams. When it supports more tx streams than rx streams, the first tx
packet is associated to the rx stream.
As I noted in previous commit, the sequence of parameters from incoming
packet is different between devices, time to time. It is worse idea to
replay the sequence of parameters from a device for the sequence of
packet to the other devices even if they are in the same category of
device.
Signed-off-by: Takashi Sakamoto <o-takashi@sakamocchi.jp>
Link: https://lore.kernel.org/r/20210527122611.173711-3-o-takashi@sakamocchi.jp
Signed-off-by: Takashi Iwai <tiwai@suse.de>
2021-05-27 12:26:10 +00:00
|
|
|
}
|
|
|
|
|
2023-01-07 02:32:14 +00:00
|
|
|
s->ctx_data.rx.cache_pos = cache_pos;
|
ALSA: firewire-lib: replay sequence of incoming packets for outgoing packets
ALSA IEC 61883-1/6 packet streaming engine uses pre-computed parameters
ideal for nominal sampling transfer frequency (STF) to transfer packets
to device since it was added 2011. As a result of user experience for a
decade, it is clear that the sequence is not suitable to some actual
devices. It takes the devices to generate noise, and causes any type of
discontinuity in the series of packet transferred from the device. It's
required for the engine to transfer packets according to effective STF.
The effective STF is given by media clock recovered by the sequence of
packet transferred from the target device. In the previous commit, the
sequence is already cached. The media clock recovery can be achieved by
analyzing the sequence.
In technological world, many ideas are proposed for media clock recovery.
However, the small part of them could be actually adopted in our case
since floating point arithmetic is not mostly available in Linux kernel
land.
This commit adopts the simple way from them; sequence replay, which means
that the sequence of parameters from incoming packet is used as is to
transfer outgoing packets. The media clock is not computed internally,
but the sequence of outgoing packet superficially looks to be generated by
the media clock.
The association between source and destination is decided when starting
AMDTP domain. When the target device supports a pair of isochronous packet
streams, the tx stream is source and the rx stream is destination. When it
supports two pair of streams, each of tx stream is associated to
corresponding rx stream in its order. When it supports less number of tx
streams than rx streams, the fist tx stream is selected for all of rx
streams. When it supports more tx streams than rx streams, the first tx
packet is associated to the rx stream.
As I noted in previous commit, the sequence of parameters from incoming
packet is different between devices, time to time. It is worse idea to
replay the sequence of parameters from a device for the sequence of
packet to the other devices even if they are in the same category of
device.
Signed-off-by: Takashi Sakamoto <o-takashi@sakamocchi.jp>
Link: https://lore.kernel.org/r/20210527122611.173711-3-o-takashi@sakamocchi.jp
Signed-off-by: Takashi Iwai <tiwai@suse.de>
2021-05-27 12:26:10 +00:00
|
|
|
}
|
|
|
|
|
2023-01-07 02:32:12 +00:00
|
|
|
static void pool_seq_descs(struct amdtp_stream *s, struct seq_desc *descs, unsigned int size,
|
|
|
|
unsigned int pos, unsigned int count)
|
ALSA: firewire-lib: replay sequence of incoming packets for outgoing packets
ALSA IEC 61883-1/6 packet streaming engine uses pre-computed parameters
ideal for nominal sampling transfer frequency (STF) to transfer packets
to device since it was added 2011. As a result of user experience for a
decade, it is clear that the sequence is not suitable to some actual
devices. It takes the devices to generate noise, and causes any type of
discontinuity in the series of packet transferred from the device. It's
required for the engine to transfer packets according to effective STF.
The effective STF is given by media clock recovered by the sequence of
packet transferred from the target device. In the previous commit, the
sequence is already cached. The media clock recovery can be achieved by
analyzing the sequence.
In technological world, many ideas are proposed for media clock recovery.
However, the small part of them could be actually adopted in our case
since floating point arithmetic is not mostly available in Linux kernel
land.
This commit adopts the simple way from them; sequence replay, which means
that the sequence of parameters from incoming packet is used as is to
transfer outgoing packets. The media clock is not computed internally,
but the sequence of outgoing packet superficially looks to be generated by
the media clock.
The association between source and destination is decided when starting
AMDTP domain. When the target device supports a pair of isochronous packet
streams, the tx stream is source and the rx stream is destination. When it
supports two pair of streams, each of tx stream is associated to
corresponding rx stream in its order. When it supports less number of tx
streams than rx streams, the fist tx stream is selected for all of rx
streams. When it supports more tx streams than rx streams, the first tx
packet is associated to the rx stream.
As I noted in previous commit, the sequence of parameters from incoming
packet is different between devices, time to time. It is worse idea to
replay the sequence of parameters from a device for the sequence of
packet to the other devices even if they are in the same category of
device.
Signed-off-by: Takashi Sakamoto <o-takashi@sakamocchi.jp>
Link: https://lore.kernel.org/r/20210527122611.173711-3-o-takashi@sakamocchi.jp
Signed-off-by: Takashi Iwai <tiwai@suse.de>
2021-05-27 12:26:10 +00:00
|
|
|
{
|
|
|
|
struct amdtp_domain *d = s->domain;
|
2023-01-07 02:32:11 +00:00
|
|
|
void (*pool_seq_descs)(struct amdtp_stream *s, struct seq_desc *descs, unsigned int size,
|
|
|
|
unsigned int pos, unsigned int count);
|
ALSA: firewire-lib: replay sequence of incoming packets for outgoing packets
ALSA IEC 61883-1/6 packet streaming engine uses pre-computed parameters
ideal for nominal sampling transfer frequency (STF) to transfer packets
to device since it was added 2011. As a result of user experience for a
decade, it is clear that the sequence is not suitable to some actual
devices. It takes the devices to generate noise, and causes any type of
discontinuity in the series of packet transferred from the device. It's
required for the engine to transfer packets according to effective STF.
The effective STF is given by media clock recovered by the sequence of
packet transferred from the target device. In the previous commit, the
sequence is already cached. The media clock recovery can be achieved by
analyzing the sequence.
In technological world, many ideas are proposed for media clock recovery.
However, the small part of them could be actually adopted in our case
since floating point arithmetic is not mostly available in Linux kernel
land.
This commit adopts the simple way from them; sequence replay, which means
that the sequence of parameters from incoming packet is used as is to
transfer outgoing packets. The media clock is not computed internally,
but the sequence of outgoing packet superficially looks to be generated by
the media clock.
The association between source and destination is decided when starting
AMDTP domain. When the target device supports a pair of isochronous packet
streams, the tx stream is source and the rx stream is destination. When it
supports two pair of streams, each of tx stream is associated to
corresponding rx stream in its order. When it supports less number of tx
streams than rx streams, the fist tx stream is selected for all of rx
streams. When it supports more tx streams than rx streams, the first tx
packet is associated to the rx stream.
As I noted in previous commit, the sequence of parameters from incoming
packet is different between devices, time to time. It is worse idea to
replay the sequence of parameters from a device for the sequence of
packet to the other devices even if they are in the same category of
device.
Signed-off-by: Takashi Sakamoto <o-takashi@sakamocchi.jp>
Link: https://lore.kernel.org/r/20210527122611.173711-3-o-takashi@sakamocchi.jp
Signed-off-by: Takashi Iwai <tiwai@suse.de>
2021-05-27 12:26:10 +00:00
|
|
|
|
2021-05-27 12:26:11 +00:00
|
|
|
if (!d->replay.enable || !s->ctx_data.rx.replay_target) {
|
2023-01-07 02:32:11 +00:00
|
|
|
pool_seq_descs = pool_ideal_seq_descs;
|
2021-05-27 12:26:11 +00:00
|
|
|
} else {
|
|
|
|
if (!d->replay.on_the_fly) {
|
2023-01-07 02:32:11 +00:00
|
|
|
pool_seq_descs = pool_replayed_seq;
|
2021-05-27 12:26:11 +00:00
|
|
|
} else {
|
|
|
|
struct amdtp_stream *tx = s->ctx_data.rx.replay_target;
|
|
|
|
const unsigned int cache_size = tx->ctx_data.tx.cache.size;
|
2023-01-07 02:32:14 +00:00
|
|
|
const unsigned int cache_pos = s->ctx_data.rx.cache_pos;
|
|
|
|
unsigned int cached_cycles = calculate_cached_cycle_count(tx, cache_pos);
|
2021-05-27 12:26:11 +00:00
|
|
|
|
|
|
|
if (cached_cycles > count && cached_cycles > cache_size / 2)
|
2023-01-07 02:32:11 +00:00
|
|
|
pool_seq_descs = pool_replayed_seq;
|
2021-05-27 12:26:11 +00:00
|
|
|
else
|
2023-01-07 02:32:11 +00:00
|
|
|
pool_seq_descs = pool_ideal_seq_descs;
|
2021-05-27 12:26:11 +00:00
|
|
|
}
|
|
|
|
}
|
2023-01-07 02:32:11 +00:00
|
|
|
|
|
|
|
pool_seq_descs(s, descs, size, pos, count);
|
ALSA: firewire-lib: replay sequence of incoming packets for outgoing packets
ALSA IEC 61883-1/6 packet streaming engine uses pre-computed parameters
ideal for nominal sampling transfer frequency (STF) to transfer packets
to device since it was added 2011. As a result of user experience for a
decade, it is clear that the sequence is not suitable to some actual
devices. It takes the devices to generate noise, and causes any type of
discontinuity in the series of packet transferred from the device. It's
required for the engine to transfer packets according to effective STF.
The effective STF is given by media clock recovered by the sequence of
packet transferred from the target device. In the previous commit, the
sequence is already cached. The media clock recovery can be achieved by
analyzing the sequence.
In technological world, many ideas are proposed for media clock recovery.
However, the small part of them could be actually adopted in our case
since floating point arithmetic is not mostly available in Linux kernel
land.
This commit adopts the simple way from them; sequence replay, which means
that the sequence of parameters from incoming packet is used as is to
transfer outgoing packets. The media clock is not computed internally,
but the sequence of outgoing packet superficially looks to be generated by
the media clock.
The association between source and destination is decided when starting
AMDTP domain. When the target device supports a pair of isochronous packet
streams, the tx stream is source and the rx stream is destination. When it
supports two pair of streams, each of tx stream is associated to
corresponding rx stream in its order. When it supports less number of tx
streams than rx streams, the fist tx stream is selected for all of rx
streams. When it supports more tx streams than rx streams, the first tx
packet is associated to the rx stream.
As I noted in previous commit, the sequence of parameters from incoming
packet is different between devices, time to time. It is worse idea to
replay the sequence of parameters from a device for the sequence of
packet to the other devices even if they are in the same category of
device.
Signed-off-by: Takashi Sakamoto <o-takashi@sakamocchi.jp>
Link: https://lore.kernel.org/r/20210527122611.173711-3-o-takashi@sakamocchi.jp
Signed-off-by: Takashi Iwai <tiwai@suse.de>
2021-05-27 12:26:10 +00:00
|
|
|
}
|
|
|
|
|
2014-04-25 13:44:45 +00:00
|
|
|
static void update_pcm_pointers(struct amdtp_stream *s,
|
|
|
|
struct snd_pcm_substream *pcm,
|
|
|
|
unsigned int frames)
|
2014-08-29 04:40:45 +00:00
|
|
|
{
|
|
|
|
unsigned int ptr;
|
|
|
|
|
2014-04-25 13:44:45 +00:00
|
|
|
ptr = s->pcm_buffer_pointer + frames;
|
|
|
|
if (ptr >= pcm->runtime->buffer_size)
|
|
|
|
ptr -= pcm->runtime->buffer_size;
|
locking/atomics: COCCINELLE/treewide: Convert trivial ACCESS_ONCE() patterns to READ_ONCE()/WRITE_ONCE()
Please do not apply this to mainline directly, instead please re-run the
coccinelle script shown below and apply its output.
For several reasons, it is desirable to use {READ,WRITE}_ONCE() in
preference to ACCESS_ONCE(), and new code is expected to use one of the
former. So far, there's been no reason to change most existing uses of
ACCESS_ONCE(), as these aren't harmful, and changing them results in
churn.
However, for some features, the read/write distinction is critical to
correct operation. To distinguish these cases, separate read/write
accessors must be used. This patch migrates (most) remaining
ACCESS_ONCE() instances to {READ,WRITE}_ONCE(), using the following
coccinelle script:
----
// Convert trivial ACCESS_ONCE() uses to equivalent READ_ONCE() and
// WRITE_ONCE()
// $ make coccicheck COCCI=/home/mark/once.cocci SPFLAGS="--include-headers" MODE=patch
virtual patch
@ depends on patch @
expression E1, E2;
@@
- ACCESS_ONCE(E1) = E2
+ WRITE_ONCE(E1, E2)
@ depends on patch @
expression E;
@@
- ACCESS_ONCE(E)
+ READ_ONCE(E)
----
Signed-off-by: Mark Rutland <mark.rutland@arm.com>
Signed-off-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: davem@davemloft.net
Cc: linux-arch@vger.kernel.org
Cc: mpe@ellerman.id.au
Cc: shuah@kernel.org
Cc: snitzer@redhat.com
Cc: thor.thayer@linux.intel.com
Cc: tj@kernel.org
Cc: viro@zeniv.linux.org.uk
Cc: will.deacon@arm.com
Link: http://lkml.kernel.org/r/1508792849-3115-19-git-send-email-paulmck@linux.vnet.ibm.com
Signed-off-by: Ingo Molnar <mingo@kernel.org>
2017-10-23 21:07:29 +00:00
|
|
|
WRITE_ONCE(s->pcm_buffer_pointer, ptr);
|
2014-04-25 13:44:45 +00:00
|
|
|
|
|
|
|
s->pcm_period_pointer += frames;
|
|
|
|
if (s->pcm_period_pointer >= pcm->runtime->period_size) {
|
|
|
|
s->pcm_period_pointer -= pcm->runtime->period_size;
|
2021-05-27 12:32:53 +00:00
|
|
|
|
|
|
|
// The program in user process should periodically check the status of intermediate
|
|
|
|
// buffer associated to PCM substream to process PCM frames in the buffer, instead
|
|
|
|
// of receiving notification of period elapsed by poll wait.
|
2024-09-04 12:51:54 +00:00
|
|
|
//
|
|
|
|
// Use another work item for period elapsed event to prevent the following AB/BA
|
|
|
|
// deadlock:
|
|
|
|
//
|
|
|
|
// thread 1 thread 2
|
|
|
|
// ================================= =================================
|
|
|
|
// A.work item (process) pcm ioctl (process)
|
|
|
|
// v v
|
|
|
|
// process_rx_packets() B.PCM stream lock
|
|
|
|
// process_tx_packets() v
|
|
|
|
// v callbacks in snd_pcm_ops
|
|
|
|
// update_pcm_pointers() v
|
|
|
|
// snd_pcm_elapsed() fw_iso_context_flush_completions()
|
|
|
|
// snd_pcm_stream_lock_irqsave() disable_work_sync()
|
|
|
|
// v v
|
|
|
|
// wait until release of B wait until A exits
|
2024-07-30 19:53:29 +00:00
|
|
|
if (!pcm->runtime->no_period_wakeup)
|
|
|
|
queue_work(system_highpri_wq, &s->period_work);
|
2014-04-25 13:44:45 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2024-07-30 19:53:26 +00:00
|
|
|
static void pcm_period_work(struct work_struct *work)
|
|
|
|
{
|
|
|
|
struct amdtp_stream *s = container_of(work, struct amdtp_stream,
|
|
|
|
period_work);
|
|
|
|
struct snd_pcm_substream *pcm = READ_ONCE(s->pcm);
|
|
|
|
|
|
|
|
if (pcm)
|
|
|
|
snd_pcm_period_elapsed(pcm);
|
|
|
|
}
|
|
|
|
|
2019-10-17 15:54:22 +00:00
|
|
|
static int queue_packet(struct amdtp_stream *s, struct fw_iso_packet *params,
|
|
|
|
bool sched_irq)
|
2014-04-25 13:44:45 +00:00
|
|
|
{
|
2019-05-23 15:14:39 +00:00
|
|
|
int err;
|
2019-05-21 14:57:36 +00:00
|
|
|
|
2019-10-17 15:54:22 +00:00
|
|
|
params->interrupt = sched_irq;
|
2019-05-23 15:14:39 +00:00
|
|
|
params->tag = s->tag;
|
|
|
|
params->sy = 0;
|
2019-05-21 14:57:36 +00:00
|
|
|
|
2019-05-23 15:14:39 +00:00
|
|
|
err = fw_iso_context_queue(s->context, params, &s->buffer.iso_buffer,
|
2014-04-25 13:44:45 +00:00
|
|
|
s->buffer.packets[s->packet_index].offset);
|
|
|
|
if (err < 0) {
|
|
|
|
dev_err(&s->unit->device, "queueing error: %d\n", err);
|
|
|
|
goto end;
|
|
|
|
}
|
|
|
|
|
2019-10-17 15:54:13 +00:00
|
|
|
if (++s->packet_index >= s->queue_size)
|
2014-04-25 13:44:45 +00:00
|
|
|
s->packet_index = 0;
|
|
|
|
end:
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int queue_out_packet(struct amdtp_stream *s,
|
2019-10-17 15:54:22 +00:00
|
|
|
struct fw_iso_packet *params, bool sched_irq)
|
2014-04-25 13:44:45 +00:00
|
|
|
{
|
2019-05-23 15:14:40 +00:00
|
|
|
params->skip =
|
|
|
|
!!(params->header_length == 0 && params->payload_length == 0);
|
2019-10-17 15:54:22 +00:00
|
|
|
return queue_packet(s, params, sched_irq);
|
2014-04-25 13:44:45 +00:00
|
|
|
}
|
|
|
|
|
2019-05-23 15:14:39 +00:00
|
|
|
static inline int queue_in_packet(struct amdtp_stream *s,
|
2019-10-18 06:19:10 +00:00
|
|
|
struct fw_iso_packet *params)
|
2014-04-25 13:44:46 +00:00
|
|
|
{
|
2019-05-23 15:14:39 +00:00
|
|
|
// Queue one packet for IR context.
|
|
|
|
params->header_length = s->ctx_data.tx.ctx_header_size;
|
|
|
|
params->payload_length = s->ctx_data.tx.max_ctx_payload_length;
|
|
|
|
params->skip = false;
|
2019-10-18 06:19:10 +00:00
|
|
|
return queue_packet(s, params, false);
|
2014-04-25 13:44:46 +00:00
|
|
|
}
|
|
|
|
|
2019-05-23 15:14:37 +00:00
|
|
|
static void generate_cip_header(struct amdtp_stream *s, __be32 cip_header[2],
|
2019-07-22 03:36:54 +00:00
|
|
|
unsigned int data_block_counter, unsigned int syt)
|
2019-05-23 15:14:37 +00:00
|
|
|
{
|
|
|
|
cip_header[0] = cpu_to_be32(READ_ONCE(s->source_node_id_field) |
|
|
|
|
(s->data_block_quadlets << CIP_DBS_SHIFT) |
|
|
|
|
((s->sph << CIP_SPH_SHIFT) & CIP_SPH_MASK) |
|
2019-07-22 03:36:54 +00:00
|
|
|
data_block_counter);
|
2019-05-23 15:14:37 +00:00
|
|
|
cip_header[1] = cpu_to_be32(CIP_EOH |
|
|
|
|
((s->fmt << CIP_FMT_SHIFT) & CIP_FMT_MASK) |
|
|
|
|
((s->ctx_data.rx.fdf << CIP_FDF_SHIFT) & CIP_FDF_MASK) |
|
|
|
|
(syt & CIP_SYT_MASK));
|
|
|
|
}
|
|
|
|
|
2019-05-24 09:03:41 +00:00
|
|
|
static void build_it_pkt_header(struct amdtp_stream *s, unsigned int cycle,
|
2021-05-20 04:01:49 +00:00
|
|
|
struct fw_iso_packet *params, unsigned int header_length,
|
2019-07-22 03:36:54 +00:00
|
|
|
unsigned int data_blocks,
|
|
|
|
unsigned int data_block_counter,
|
2023-01-09 21:32:31 +00:00
|
|
|
unsigned int syt, unsigned int index, u32 curr_cycle_time)
|
2011-03-15 06:53:21 +00:00
|
|
|
{
|
2019-07-07 12:07:53 +00:00
|
|
|
unsigned int payload_length;
|
2019-05-23 15:14:38 +00:00
|
|
|
__be32 *cip_header;
|
2015-09-19 02:21:52 +00:00
|
|
|
|
2019-07-07 12:07:53 +00:00
|
|
|
payload_length = data_blocks * sizeof(__be32) * s->data_block_quadlets;
|
|
|
|
params->payload_length = payload_length;
|
|
|
|
|
2021-05-20 04:01:49 +00:00
|
|
|
if (header_length > 0) {
|
2019-05-24 09:03:41 +00:00
|
|
|
cip_header = (__be32 *)params->header;
|
2019-07-22 03:36:54 +00:00
|
|
|
generate_cip_header(s, cip_header, data_block_counter, syt);
|
2021-05-20 04:01:49 +00:00
|
|
|
params->header_length = header_length;
|
2019-05-23 15:14:40 +00:00
|
|
|
} else {
|
|
|
|
cip_header = NULL;
|
|
|
|
}
|
2011-03-15 06:53:21 +00:00
|
|
|
|
2021-05-20 04:01:49 +00:00
|
|
|
trace_amdtp_packet(s, cycle, cip_header, payload_length + header_length, data_blocks,
|
2023-01-09 21:32:31 +00:00
|
|
|
data_block_counter, s->packet_index, index, curr_cycle_time);
|
ALSA: firewire-lib: add no-header packet processing
As long as investigating Fireface 400, IEC 61883-1/6 is not applied to
its packet streaming protocol. Remarks of the specific protocol are:
* Each packet doesn't include CIP headers.
* 64,0 and 128,0 kHz are supported.
* The device doesn't necessarily transmit 8,000 packets per second.
* 0, 1, 2, 3 are used as tag for rx isochronous packets, however 0 is
used for tx isochronous packets.
On the other hand, there's a common feature. The number of data blocks
transferred in a second is the same as sampling transmission frequency.
Current ALSA IEC 61883-1/6 engine already has a method to calculate it and
this driver can utilize it for rx packets, as well as tx packets.
This commit adds support for the transferring protocol. CIP_NO_HEADERS
flag is newly added. When this flag is set:
* Both of 0 (without CIP header) and 1 (with CIP header) are used as tag
to handle incoming isochronous packet.
* 0 (without CIP header) is used as tag to transfer outgoing isochronous
packet.
* Skip CIP header evaluation.
* Use unique way to calculate the quadlets of isochronous packet payload.
In ALSA PCM interface, 128.0 kHz is not supported, and the ALSA
IEC 61883-1/6 engine doesn't support 64.0 kHz. These modes are dropped.
The sequence of rx packet has a remarkable quirk about tag. This will be
described in later commits.
Signed-off-by: Takashi Sakamoto <o-takashi@sakamocchi.jp>
Signed-off-by: Takashi Iwai <tiwai@suse.de>
2017-03-31 13:06:07 +00:00
|
|
|
}
|
|
|
|
|
2019-05-22 14:17:06 +00:00
|
|
|
static int check_cip_header(struct amdtp_stream *s, const __be32 *buf,
|
|
|
|
unsigned int payload_length,
|
2019-07-22 03:36:55 +00:00
|
|
|
unsigned int *data_blocks,
|
|
|
|
unsigned int *data_block_counter, unsigned int *syt)
|
2014-04-25 13:44:46 +00:00
|
|
|
{
|
|
|
|
u32 cip_header[2];
|
2019-05-22 14:17:06 +00:00
|
|
|
unsigned int sph;
|
|
|
|
unsigned int fmt;
|
|
|
|
unsigned int fdf;
|
2019-07-22 03:36:55 +00:00
|
|
|
unsigned int dbc;
|
2014-04-25 13:45:04 +00:00
|
|
|
bool lost;
|
2014-04-25 13:44:46 +00:00
|
|
|
|
2019-05-22 14:17:06 +00:00
|
|
|
cip_header[0] = be32_to_cpu(buf[0]);
|
|
|
|
cip_header[1] = be32_to_cpu(buf[1]);
|
2014-04-25 13:44:46 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* This module supports 'Two-quadlet CIP header with SYT field'.
|
2014-04-25 13:44:50 +00:00
|
|
|
* For convenience, also check FMT field is AM824 or not.
|
2014-04-25 13:44:46 +00:00
|
|
|
*/
|
2017-03-22 12:30:27 +00:00
|
|
|
if ((((cip_header[0] & CIP_EOH_MASK) == CIP_EOH) ||
|
|
|
|
((cip_header[1] & CIP_EOH_MASK) != CIP_EOH)) &&
|
|
|
|
(!(s->flags & CIP_HEADER_WITHOUT_EOH))) {
|
2014-04-25 13:44:46 +00:00
|
|
|
dev_info_ratelimited(&s->unit->device,
|
|
|
|
"Invalid CIP header for AMDTP: %08X:%08X\n",
|
|
|
|
cip_header[0], cip_header[1]);
|
2019-05-22 14:17:06 +00:00
|
|
|
return -EAGAIN;
|
2014-04-25 13:44:46 +00:00
|
|
|
}
|
|
|
|
|
2015-09-19 02:21:53 +00:00
|
|
|
/* Check valid protocol or not. */
|
2017-03-22 12:30:16 +00:00
|
|
|
sph = (cip_header[0] & CIP_SPH_MASK) >> CIP_SPH_SHIFT;
|
2015-09-19 02:21:53 +00:00
|
|
|
fmt = (cip_header[1] & CIP_FMT_MASK) >> CIP_FMT_SHIFT;
|
2017-03-22 12:30:16 +00:00
|
|
|
if (sph != s->sph || fmt != s->fmt) {
|
2015-10-11 13:33:50 +00:00
|
|
|
dev_info_ratelimited(&s->unit->device,
|
|
|
|
"Detect unexpected protocol: %08x %08x\n",
|
|
|
|
cip_header[0], cip_header[1]);
|
2019-05-22 14:17:06 +00:00
|
|
|
return -EAGAIN;
|
2015-09-19 02:21:53 +00:00
|
|
|
}
|
|
|
|
|
2014-04-25 13:44:46 +00:00
|
|
|
/* Calculate data blocks */
|
2015-09-19 02:21:53 +00:00
|
|
|
fdf = (cip_header[1] & CIP_FDF_MASK) >> CIP_FDF_SHIFT;
|
2021-05-18 13:00:45 +00:00
|
|
|
if (payload_length == 0 || (fmt == CIP_FMT_AM && fdf == AMDTP_FDF_NO_DATA)) {
|
2019-05-22 14:17:06 +00:00
|
|
|
*data_blocks = 0;
|
2014-04-25 13:44:46 +00:00
|
|
|
} else {
|
2019-05-22 14:17:06 +00:00
|
|
|
unsigned int data_block_quadlets =
|
|
|
|
(cip_header[0] & CIP_DBS_MASK) >> CIP_DBS_SHIFT;
|
2014-04-25 13:44:46 +00:00
|
|
|
/* avoid division by zero */
|
|
|
|
if (data_block_quadlets == 0) {
|
2015-05-22 14:21:13 +00:00
|
|
|
dev_err(&s->unit->device,
|
2014-04-25 13:44:46 +00:00
|
|
|
"Detect invalid value in dbs field: %08X\n",
|
|
|
|
cip_header[0]);
|
2015-05-22 14:21:14 +00:00
|
|
|
return -EPROTO;
|
2014-04-25 13:44:46 +00:00
|
|
|
}
|
2014-04-25 13:45:05 +00:00
|
|
|
if (s->flags & CIP_WRONG_DBS)
|
|
|
|
data_block_quadlets = s->data_block_quadlets;
|
2014-04-25 13:44:46 +00:00
|
|
|
|
2021-05-18 13:00:45 +00:00
|
|
|
*data_blocks = payload_length / sizeof(__be32) / data_block_quadlets;
|
2014-04-25 13:44:46 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/* Check data block counter continuity */
|
2019-07-22 03:36:55 +00:00
|
|
|
dbc = cip_header[0] & CIP_DBC_MASK;
|
2019-05-22 14:17:06 +00:00
|
|
|
if (*data_blocks == 0 && (s->flags & CIP_EMPTY_HAS_WRONG_DBC) &&
|
2019-07-22 03:36:55 +00:00
|
|
|
*data_block_counter != UINT_MAX)
|
|
|
|
dbc = *data_block_counter;
|
2014-04-25 13:45:27 +00:00
|
|
|
|
2019-07-22 03:36:55 +00:00
|
|
|
if ((dbc == 0x00 && (s->flags & CIP_SKIP_DBC_ZERO_CHECK)) ||
|
|
|
|
*data_block_counter == UINT_MAX) {
|
2014-04-25 13:45:07 +00:00
|
|
|
lost = false;
|
|
|
|
} else if (!(s->flags & CIP_DBC_IS_END_EVENT)) {
|
2019-07-22 03:36:55 +00:00
|
|
|
lost = dbc != *data_block_counter;
|
2014-04-25 13:45:06 +00:00
|
|
|
} else {
|
2019-05-22 14:17:06 +00:00
|
|
|
unsigned int dbc_interval;
|
|
|
|
|
2024-02-18 07:41:27 +00:00
|
|
|
if (!(s->flags & CIP_DBC_IS_PAYLOAD_QUADLETS)) {
|
|
|
|
if (*data_blocks > 0 && s->ctx_data.tx.dbc_interval > 0)
|
|
|
|
dbc_interval = s->ctx_data.tx.dbc_interval;
|
|
|
|
else
|
|
|
|
dbc_interval = *data_blocks;
|
|
|
|
} else {
|
|
|
|
dbc_interval = payload_length / sizeof(__be32);
|
|
|
|
}
|
2014-04-25 13:45:06 +00:00
|
|
|
|
2019-07-22 03:36:55 +00:00
|
|
|
lost = dbc != ((*data_block_counter + dbc_interval) & 0xff);
|
2014-04-25 13:45:06 +00:00
|
|
|
}
|
2014-04-25 13:45:04 +00:00
|
|
|
|
|
|
|
if (lost) {
|
2015-05-22 14:21:13 +00:00
|
|
|
dev_err(&s->unit->device,
|
|
|
|
"Detect discontinuity of CIP: %02X %02X\n",
|
2019-07-22 03:36:55 +00:00
|
|
|
*data_block_counter, dbc);
|
2015-05-22 14:00:52 +00:00
|
|
|
return -EIO;
|
2014-04-25 13:44:46 +00:00
|
|
|
}
|
|
|
|
|
2019-07-22 03:37:00 +00:00
|
|
|
*data_block_counter = dbc;
|
|
|
|
|
2021-05-22 01:32:58 +00:00
|
|
|
if (!(s->flags & CIP_UNAWARE_SYT))
|
|
|
|
*syt = cip_header[1] & CIP_SYT_MASK;
|
2014-04-25 13:44:46 +00:00
|
|
|
|
2019-05-22 14:17:06 +00:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2019-05-24 09:03:42 +00:00
|
|
|
static int parse_ir_ctx_header(struct amdtp_stream *s, unsigned int cycle,
|
|
|
|
const __be32 *ctx_header,
|
2019-07-22 03:36:55 +00:00
|
|
|
unsigned int *data_blocks,
|
|
|
|
unsigned int *data_block_counter,
|
2023-01-09 21:32:31 +00:00
|
|
|
unsigned int *syt, unsigned int packet_index, unsigned int index,
|
|
|
|
u32 curr_cycle_time)
|
2019-05-22 14:17:06 +00:00
|
|
|
{
|
2021-05-18 13:00:44 +00:00
|
|
|
unsigned int payload_length;
|
2019-05-22 14:17:07 +00:00
|
|
|
const __be32 *cip_header;
|
2021-05-13 12:56:50 +00:00
|
|
|
unsigned int cip_header_size;
|
2019-05-22 14:17:06 +00:00
|
|
|
|
2021-05-18 13:00:44 +00:00
|
|
|
payload_length = be32_to_cpu(ctx_header[0]) >> ISO_DATA_LENGTH_SHIFT;
|
2021-05-13 12:56:50 +00:00
|
|
|
|
|
|
|
if (!(s->flags & CIP_NO_HEADER))
|
2021-05-20 04:01:47 +00:00
|
|
|
cip_header_size = CIP_HEADER_SIZE;
|
2021-05-13 12:56:50 +00:00
|
|
|
else
|
|
|
|
cip_header_size = 0;
|
|
|
|
|
2021-05-18 13:00:44 +00:00
|
|
|
if (payload_length > cip_header_size + s->ctx_data.tx.max_ctx_payload_length) {
|
2019-05-22 14:17:06 +00:00
|
|
|
dev_err(&s->unit->device,
|
|
|
|
"Detect jumbo payload: %04x %04x\n",
|
2021-05-18 13:00:44 +00:00
|
|
|
payload_length, cip_header_size + s->ctx_data.tx.max_ctx_payload_length);
|
2019-05-22 14:17:06 +00:00
|
|
|
return -EIO;
|
|
|
|
}
|
|
|
|
|
2021-05-13 12:56:50 +00:00
|
|
|
if (cip_header_size > 0) {
|
2021-05-18 13:00:44 +00:00
|
|
|
if (payload_length >= cip_header_size) {
|
2021-05-20 13:04:09 +00:00
|
|
|
int err;
|
|
|
|
|
2021-05-20 04:01:47 +00:00
|
|
|
cip_header = ctx_header + IR_CTX_HEADER_DEFAULT_QUADLETS;
|
2021-05-18 13:00:45 +00:00
|
|
|
err = check_cip_header(s, cip_header, payload_length - cip_header_size,
|
|
|
|
data_blocks, data_block_counter, syt);
|
2021-05-18 13:00:41 +00:00
|
|
|
if (err < 0)
|
|
|
|
return err;
|
|
|
|
} else {
|
|
|
|
// Handle the cycle so that empty packet arrives.
|
|
|
|
cip_header = NULL;
|
|
|
|
*data_blocks = 0;
|
|
|
|
*syt = 0;
|
|
|
|
}
|
2019-05-22 14:17:08 +00:00
|
|
|
} else {
|
|
|
|
cip_header = NULL;
|
2021-05-18 13:00:44 +00:00
|
|
|
*data_blocks = payload_length / sizeof(__be32) / s->data_block_quadlets;
|
2019-05-24 09:03:42 +00:00
|
|
|
*syt = 0;
|
2019-07-07 12:07:54 +00:00
|
|
|
|
2019-07-22 03:36:55 +00:00
|
|
|
if (*data_block_counter == UINT_MAX)
|
|
|
|
*data_block_counter = 0;
|
2019-05-22 14:17:06 +00:00
|
|
|
}
|
|
|
|
|
2021-05-18 13:00:44 +00:00
|
|
|
trace_amdtp_packet(s, cycle, cip_header, payload_length, *data_blocks,
|
2023-01-09 21:32:31 +00:00
|
|
|
*data_block_counter, packet_index, index, curr_cycle_time);
|
2019-05-22 14:17:06 +00:00
|
|
|
|
2021-05-20 13:04:09 +00:00
|
|
|
return 0;
|
2014-04-25 13:44:46 +00:00
|
|
|
}
|
|
|
|
|
2019-05-21 14:57:37 +00:00
|
|
|
// In CYCLE_TIMER register of IEEE 1394, 7 bits are used to represent second. On
|
|
|
|
// the other hand, in DMA descriptors of 1394 OHCI, 3 bits are used to represent
|
|
|
|
// it. Thus, via Linux firewire subsystem, we can get the 3 bits for second.
|
2023-01-10 13:49:33 +00:00
|
|
|
static inline u32 compute_ohci_iso_ctx_cycle_count(u32 tstamp)
|
|
|
|
{
|
|
|
|
return (((tstamp >> 13) & 0x07) * CYCLES_PER_SECOND) + (tstamp & 0x1fff);
|
|
|
|
}
|
|
|
|
|
2021-05-18 13:00:43 +00:00
|
|
|
static inline u32 compute_ohci_cycle_count(__be32 ctx_header_tstamp)
|
ALSA: firewire-lib: compute the value of second field in cycle count for IT context
In callback function of isochronous context, u32 variable is passed for
cycle count. The value of this variable comes from DMA descriptors of 1394
Open Host Controller Interface (1394 OHCI). In the specification, DMA
descriptors transport lower 3 bits for second field and full cycle field in
16 bits field, therefore 16 bits of the u32 variable are available. The
value for second is modulo 8, and the value for cycle is modulo 8,000.
Currently, ALSA firewire-lib module don't use the value of the second
field, because the value is useless to calculate presentation timestamp in
IEC 61883-6. However, the value may be useful for debugging. In later
commit, it will be printed with the other parameters for debugging.
This commit makes this module to handle the whole cycle count including
second. The value is calculated by cycle unit. The existed code is already
written with ignoring the value of second, thus this commit causes no
issues.
Signed-off-by: Takashi Sakamoto <o-takashi@sakamocchi.jp>
Signed-off-by: Takashi Iwai <tiwai@suse.de>
2016-05-09 12:12:44 +00:00
|
|
|
{
|
2019-05-21 14:57:37 +00:00
|
|
|
u32 tstamp = be32_to_cpu(ctx_header_tstamp) & HEADER_TSTAMP_MASK;
|
2023-01-10 13:49:33 +00:00
|
|
|
return compute_ohci_iso_ctx_cycle_count(tstamp);
|
ALSA: firewire-lib: compute the value of second field in cycle count for IT context
In callback function of isochronous context, u32 variable is passed for
cycle count. The value of this variable comes from DMA descriptors of 1394
Open Host Controller Interface (1394 OHCI). In the specification, DMA
descriptors transport lower 3 bits for second field and full cycle field in
16 bits field, therefore 16 bits of the u32 variable are available. The
value for second is modulo 8, and the value for cycle is modulo 8,000.
Currently, ALSA firewire-lib module don't use the value of the second
field, because the value is useless to calculate presentation timestamp in
IEC 61883-6. However, the value may be useful for debugging. In later
commit, it will be printed with the other parameters for debugging.
This commit makes this module to handle the whole cycle count including
second. The value is calculated by cycle unit. The existed code is already
written with ignoring the value of second, thus this commit causes no
issues.
Signed-off-by: Takashi Sakamoto <o-takashi@sakamocchi.jp>
Signed-off-by: Takashi Iwai <tiwai@suse.de>
2016-05-09 12:12:44 +00:00
|
|
|
}
|
|
|
|
|
2021-05-18 13:00:43 +00:00
|
|
|
static inline u32 increment_ohci_cycle_count(u32 cycle, unsigned int addend)
|
ALSA: firewire-lib: compute the value of second field in cycle count for IT context
In callback function of isochronous context, u32 variable is passed for
cycle count. The value of this variable comes from DMA descriptors of 1394
Open Host Controller Interface (1394 OHCI). In the specification, DMA
descriptors transport lower 3 bits for second field and full cycle field in
16 bits field, therefore 16 bits of the u32 variable are available. The
value for second is modulo 8, and the value for cycle is modulo 8,000.
Currently, ALSA firewire-lib module don't use the value of the second
field, because the value is useless to calculate presentation timestamp in
IEC 61883-6. However, the value may be useful for debugging. In later
commit, it will be printed with the other parameters for debugging.
This commit makes this module to handle the whole cycle count including
second. The value is calculated by cycle unit. The existed code is already
written with ignoring the value of second, thus this commit causes no
issues.
Signed-off-by: Takashi Sakamoto <o-takashi@sakamocchi.jp>
Signed-off-by: Takashi Iwai <tiwai@suse.de>
2016-05-09 12:12:44 +00:00
|
|
|
{
|
|
|
|
cycle += addend;
|
2021-05-18 13:00:43 +00:00
|
|
|
if (cycle >= OHCI_SECOND_MODULUS * CYCLES_PER_SECOND)
|
|
|
|
cycle -= OHCI_SECOND_MODULUS * CYCLES_PER_SECOND;
|
ALSA: firewire-lib: compute the value of second field in cycle count for IT context
In callback function of isochronous context, u32 variable is passed for
cycle count. The value of this variable comes from DMA descriptors of 1394
Open Host Controller Interface (1394 OHCI). In the specification, DMA
descriptors transport lower 3 bits for second field and full cycle field in
16 bits field, therefore 16 bits of the u32 variable are available. The
value for second is modulo 8, and the value for cycle is modulo 8,000.
Currently, ALSA firewire-lib module don't use the value of the second
field, because the value is useless to calculate presentation timestamp in
IEC 61883-6. However, the value may be useful for debugging. In later
commit, it will be printed with the other parameters for debugging.
This commit makes this module to handle the whole cycle count including
second. The value is calculated by cycle unit. The existed code is already
written with ignoring the value of second, thus this commit causes no
issues.
Signed-off-by: Takashi Sakamoto <o-takashi@sakamocchi.jp>
Signed-off-by: Takashi Iwai <tiwai@suse.de>
2016-05-09 12:12:44 +00:00
|
|
|
return cycle;
|
|
|
|
}
|
|
|
|
|
2023-01-10 13:49:33 +00:00
|
|
|
static inline u32 decrement_ohci_cycle_count(u32 minuend, u32 subtrahend)
|
|
|
|
{
|
|
|
|
if (minuend < subtrahend)
|
|
|
|
minuend += OHCI_SECOND_MODULUS * CYCLES_PER_SECOND;
|
|
|
|
|
|
|
|
return minuend - subtrahend;
|
|
|
|
}
|
|
|
|
|
2021-05-18 13:00:46 +00:00
|
|
|
static int compare_ohci_cycle_count(u32 lval, u32 rval)
|
|
|
|
{
|
|
|
|
if (lval == rval)
|
|
|
|
return 0;
|
|
|
|
else if (lval < rval && rval - lval < OHCI_SECOND_MODULUS * CYCLES_PER_SECOND / 2)
|
|
|
|
return -1;
|
|
|
|
else
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2019-05-21 14:57:37 +00:00
|
|
|
// Align to actual cycle count for the packet which is going to be scheduled.
|
2019-10-17 15:54:13 +00:00
|
|
|
// This module queued the same number of isochronous cycle as the size of queue
|
|
|
|
// to kip isochronous cycle, therefore it's OK to just increment the cycle by
|
|
|
|
// the size of queue for scheduled cycle.
|
2021-05-18 13:00:43 +00:00
|
|
|
static inline u32 compute_ohci_it_cycle(const __be32 ctx_header_tstamp,
|
|
|
|
unsigned int queue_size)
|
2019-05-21 14:57:37 +00:00
|
|
|
{
|
2021-05-18 13:00:43 +00:00
|
|
|
u32 cycle = compute_ohci_cycle_count(ctx_header_tstamp);
|
|
|
|
return increment_ohci_cycle_count(cycle, queue_size);
|
2019-05-21 14:57:37 +00:00
|
|
|
}
|
|
|
|
|
2023-01-09 02:17:36 +00:00
|
|
|
static int generate_tx_packet_descs(struct amdtp_stream *s, struct pkt_desc *desc,
|
2023-01-07 02:32:13 +00:00
|
|
|
const __be32 *ctx_header, unsigned int packet_count,
|
|
|
|
unsigned int *desc_count)
|
2019-07-22 03:37:00 +00:00
|
|
|
{
|
2021-05-18 13:00:46 +00:00
|
|
|
unsigned int next_cycle = s->next_cycle;
|
2019-07-22 03:37:00 +00:00
|
|
|
unsigned int dbc = s->data_block_counter;
|
2021-05-13 12:56:52 +00:00
|
|
|
unsigned int packet_index = s->packet_index;
|
|
|
|
unsigned int queue_size = s->queue_size;
|
2023-01-30 14:15:32 +00:00
|
|
|
u32 curr_cycle_time = 0;
|
2019-07-22 03:37:00 +00:00
|
|
|
int i;
|
|
|
|
int err;
|
|
|
|
|
2023-01-09 21:32:31 +00:00
|
|
|
if (trace_amdtp_packet_enabled())
|
|
|
|
(void)fw_card_read_cycle_time(fw_parent_device(s->unit)->card, &curr_cycle_time);
|
|
|
|
|
2021-05-18 13:00:47 +00:00
|
|
|
*desc_count = 0;
|
2023-01-07 02:32:13 +00:00
|
|
|
for (i = 0; i < packet_count; ++i) {
|
2019-07-22 03:37:00 +00:00
|
|
|
unsigned int cycle;
|
2021-05-18 13:00:46 +00:00
|
|
|
bool lost;
|
2019-07-22 03:37:00 +00:00
|
|
|
unsigned int data_blocks;
|
|
|
|
unsigned int syt;
|
|
|
|
|
2021-05-18 13:00:43 +00:00
|
|
|
cycle = compute_ohci_cycle_count(ctx_header[1]);
|
2021-05-18 13:00:46 +00:00
|
|
|
lost = (next_cycle != cycle);
|
|
|
|
if (lost) {
|
|
|
|
if (s->flags & CIP_NO_HEADER) {
|
|
|
|
// Fireface skips transmission just for an isoc cycle corresponding
|
|
|
|
// to empty packet.
|
2021-05-18 13:00:47 +00:00
|
|
|
unsigned int prev_cycle = next_cycle;
|
|
|
|
|
2021-05-18 13:00:46 +00:00
|
|
|
next_cycle = increment_ohci_cycle_count(next_cycle, 1);
|
|
|
|
lost = (next_cycle != cycle);
|
2021-05-18 13:00:47 +00:00
|
|
|
if (!lost) {
|
|
|
|
// Prepare a description for the skipped cycle for
|
|
|
|
// sequence replay.
|
|
|
|
desc->cycle = prev_cycle;
|
|
|
|
desc->syt = 0;
|
|
|
|
desc->data_blocks = 0;
|
|
|
|
desc->data_block_counter = dbc;
|
|
|
|
desc->ctx_payload = NULL;
|
2023-01-09 02:17:36 +00:00
|
|
|
desc = amdtp_stream_next_packet_desc(s, desc);
|
2021-05-18 13:00:47 +00:00
|
|
|
++(*desc_count);
|
|
|
|
}
|
2021-05-18 13:00:46 +00:00
|
|
|
} else if (s->flags & CIP_JUMBO_PAYLOAD) {
|
|
|
|
// OXFW970 skips transmission for several isoc cycles during
|
2021-05-18 13:00:47 +00:00
|
|
|
// asynchronous transaction. The sequence replay is impossible due
|
|
|
|
// to the reason.
|
2021-05-18 13:00:46 +00:00
|
|
|
unsigned int safe_cycle = increment_ohci_cycle_count(next_cycle,
|
|
|
|
IR_JUMBO_PAYLOAD_MAX_SKIP_CYCLES);
|
2024-02-18 03:30:26 +00:00
|
|
|
lost = (compare_ohci_cycle_count(safe_cycle, cycle) < 0);
|
2021-05-18 13:00:46 +00:00
|
|
|
}
|
|
|
|
if (lost) {
|
|
|
|
dev_err(&s->unit->device, "Detect discontinuity of cycle: %d %d\n",
|
|
|
|
next_cycle, cycle);
|
|
|
|
return -EIO;
|
|
|
|
}
|
|
|
|
}
|
2019-07-22 03:37:00 +00:00
|
|
|
|
2021-05-18 13:00:44 +00:00
|
|
|
err = parse_ir_ctx_header(s, cycle, ctx_header, &data_blocks, &dbc, &syt,
|
2023-01-09 21:32:31 +00:00
|
|
|
packet_index, i, curr_cycle_time);
|
2019-07-22 03:37:00 +00:00
|
|
|
if (err < 0)
|
|
|
|
return err;
|
|
|
|
|
|
|
|
desc->cycle = cycle;
|
|
|
|
desc->syt = syt;
|
|
|
|
desc->data_blocks = data_blocks;
|
|
|
|
desc->data_block_counter = dbc;
|
2021-05-13 12:56:52 +00:00
|
|
|
desc->ctx_payload = s->buffer.packets[packet_index].buffer;
|
2019-07-22 03:37:00 +00:00
|
|
|
|
|
|
|
if (!(s->flags & CIP_DBC_IS_END_EVENT))
|
|
|
|
dbc = (dbc + desc->data_blocks) & 0xff;
|
|
|
|
|
2021-05-18 13:00:46 +00:00
|
|
|
next_cycle = increment_ohci_cycle_count(next_cycle, 1);
|
2023-01-09 02:17:36 +00:00
|
|
|
desc = amdtp_stream_next_packet_desc(s, desc);
|
2021-05-18 13:00:47 +00:00
|
|
|
++(*desc_count);
|
2021-05-18 13:00:46 +00:00
|
|
|
ctx_header += s->ctx_data.tx.ctx_header_size / sizeof(*ctx_header);
|
2021-05-13 12:56:52 +00:00
|
|
|
packet_index = (packet_index + 1) % queue_size;
|
2019-07-22 03:37:00 +00:00
|
|
|
}
|
|
|
|
|
2021-05-18 13:00:46 +00:00
|
|
|
s->next_cycle = next_cycle;
|
2019-07-22 03:37:00 +00:00
|
|
|
s->data_block_counter = dbc;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2020-05-08 04:36:30 +00:00
|
|
|
static unsigned int compute_syt(unsigned int syt_offset, unsigned int cycle,
|
|
|
|
unsigned int transfer_delay)
|
|
|
|
{
|
|
|
|
unsigned int syt;
|
|
|
|
|
|
|
|
syt_offset += transfer_delay;
|
|
|
|
syt = ((cycle + syt_offset / TICKS_PER_CYCLE) << 12) |
|
|
|
|
(syt_offset % TICKS_PER_CYCLE);
|
|
|
|
return syt & CIP_SYT_MASK;
|
|
|
|
}
|
|
|
|
|
2023-01-09 02:17:36 +00:00
|
|
|
static void generate_rx_packet_descs(struct amdtp_stream *s, struct pkt_desc *desc,
|
2023-01-07 02:32:12 +00:00
|
|
|
const __be32 *ctx_header, unsigned int packet_count)
|
2019-07-22 03:36:59 +00:00
|
|
|
{
|
2023-01-07 02:32:12 +00:00
|
|
|
struct seq_desc *seq_descs = s->ctx_data.rx.seq.descs;
|
|
|
|
unsigned int seq_size = s->ctx_data.rx.seq.size;
|
|
|
|
unsigned int seq_pos = s->ctx_data.rx.seq.pos;
|
2019-07-22 03:36:59 +00:00
|
|
|
unsigned int dbc = s->data_block_counter;
|
2021-05-22 01:32:58 +00:00
|
|
|
bool aware_syt = !(s->flags & CIP_UNAWARE_SYT);
|
2019-07-22 03:36:59 +00:00
|
|
|
int i;
|
|
|
|
|
2023-01-07 02:32:12 +00:00
|
|
|
pool_seq_descs(s, seq_descs, seq_size, seq_pos, packet_count);
|
|
|
|
|
|
|
|
for (i = 0; i < packet_count; ++i) {
|
2019-10-17 15:54:13 +00:00
|
|
|
unsigned int index = (s->packet_index + i) % s->queue_size;
|
2023-01-07 02:32:12 +00:00
|
|
|
const struct seq_desc *seq = seq_descs + seq_pos;
|
2019-07-22 03:36:59 +00:00
|
|
|
|
2021-05-18 13:00:43 +00:00
|
|
|
desc->cycle = compute_ohci_it_cycle(*ctx_header, s->queue_size);
|
2020-05-08 04:36:35 +00:00
|
|
|
|
2021-05-22 01:33:03 +00:00
|
|
|
if (aware_syt && seq->syt_offset != CIP_SYT_NO_INFO)
|
|
|
|
desc->syt = compute_syt(seq->syt_offset, desc->cycle, s->transfer_delay);
|
|
|
|
else
|
2021-05-22 01:32:58 +00:00
|
|
|
desc->syt = CIP_SYT_NO_INFO;
|
|
|
|
|
2020-05-08 04:36:35 +00:00
|
|
|
desc->data_blocks = seq->data_blocks;
|
2019-07-22 03:36:59 +00:00
|
|
|
|
|
|
|
if (s->flags & CIP_DBC_IS_END_EVENT)
|
|
|
|
dbc = (dbc + desc->data_blocks) & 0xff;
|
|
|
|
|
|
|
|
desc->data_block_counter = dbc;
|
|
|
|
|
|
|
|
if (!(s->flags & CIP_DBC_IS_END_EVENT))
|
|
|
|
dbc = (dbc + desc->data_blocks) & 0xff;
|
|
|
|
|
|
|
|
desc->ctx_payload = s->buffer.packets[index].buffer;
|
|
|
|
|
2023-01-07 02:32:12 +00:00
|
|
|
seq_pos = (seq_pos + 1) % seq_size;
|
2023-01-09 02:17:36 +00:00
|
|
|
desc = amdtp_stream_next_packet_desc(s, desc);
|
2020-05-08 04:36:35 +00:00
|
|
|
|
2019-07-22 03:36:59 +00:00
|
|
|
++ctx_header;
|
|
|
|
}
|
|
|
|
|
|
|
|
s->data_block_counter = dbc;
|
2023-01-07 02:32:12 +00:00
|
|
|
s->ctx_data.rx.seq.pos = seq_pos;
|
2019-07-22 03:36:59 +00:00
|
|
|
}
|
|
|
|
|
2019-05-21 14:57:35 +00:00
|
|
|
static inline void cancel_stream(struct amdtp_stream *s)
|
|
|
|
{
|
2024-09-04 12:51:54 +00:00
|
|
|
struct work_struct *work = current_work();
|
|
|
|
|
2019-05-21 14:57:35 +00:00
|
|
|
s->packet_index = -1;
|
2024-09-04 12:51:54 +00:00
|
|
|
|
|
|
|
// Detect work items for any isochronous context. The work item for pcm_period_work()
|
|
|
|
// should be avoided since the call of snd_pcm_period_elapsed() can reach via
|
|
|
|
// snd_pcm_ops.pointer() under acquiring PCM stream(group) lock and causes dead lock at
|
|
|
|
// snd_pcm_stop_xrun().
|
|
|
|
if (work && work != &s->period_work)
|
2019-05-21 14:57:35 +00:00
|
|
|
amdtp_stream_pcm_abort(s);
|
|
|
|
WRITE_ONCE(s->pcm_buffer_pointer, SNDRV_PCM_POS_XRUN);
|
|
|
|
}
|
|
|
|
|
2023-01-10 13:49:33 +00:00
|
|
|
static snd_pcm_sframes_t compute_pcm_extra_delay(struct amdtp_stream *s,
|
|
|
|
const struct pkt_desc *desc, unsigned int count)
|
|
|
|
{
|
|
|
|
unsigned int data_block_count = 0;
|
|
|
|
u32 latest_cycle;
|
|
|
|
u32 cycle_time;
|
|
|
|
u32 curr_cycle;
|
|
|
|
u32 cycle_gap;
|
|
|
|
int i, err;
|
|
|
|
|
|
|
|
if (count == 0)
|
|
|
|
goto end;
|
|
|
|
|
|
|
|
// Forward to the latest record.
|
|
|
|
for (i = 0; i < count - 1; ++i)
|
|
|
|
desc = amdtp_stream_next_packet_desc(s, desc);
|
|
|
|
latest_cycle = desc->cycle;
|
|
|
|
|
|
|
|
err = fw_card_read_cycle_time(fw_parent_device(s->unit)->card, &cycle_time);
|
|
|
|
if (err < 0)
|
|
|
|
goto end;
|
|
|
|
|
|
|
|
// Compute cycle count with lower 3 bits of second field and cycle field like timestamp
|
|
|
|
// format of 1394 OHCI isochronous context.
|
|
|
|
curr_cycle = compute_ohci_iso_ctx_cycle_count((cycle_time >> 12) & 0x0000ffff);
|
|
|
|
|
|
|
|
if (s->direction == AMDTP_IN_STREAM) {
|
|
|
|
// NOTE: The AMDTP packet descriptor should be for the past isochronous cycle since
|
|
|
|
// it corresponds to arrived isochronous packet.
|
|
|
|
if (compare_ohci_cycle_count(latest_cycle, curr_cycle) > 0)
|
|
|
|
goto end;
|
|
|
|
cycle_gap = decrement_ohci_cycle_count(curr_cycle, latest_cycle);
|
|
|
|
|
|
|
|
// NOTE: estimate delay by recent history of arrived AMDTP packets. The estimated
|
|
|
|
// value expectedly corresponds to a few packets (0-2) since the packet arrived at
|
|
|
|
// the most recent isochronous cycle has been already processed.
|
|
|
|
for (i = 0; i < cycle_gap; ++i) {
|
|
|
|
desc = amdtp_stream_next_packet_desc(s, desc);
|
|
|
|
data_block_count += desc->data_blocks;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
// NOTE: The AMDTP packet descriptor should be for the future isochronous cycle
|
|
|
|
// since it was already scheduled.
|
|
|
|
if (compare_ohci_cycle_count(latest_cycle, curr_cycle) < 0)
|
|
|
|
goto end;
|
|
|
|
cycle_gap = decrement_ohci_cycle_count(latest_cycle, curr_cycle);
|
|
|
|
|
|
|
|
// NOTE: use history of scheduled packets.
|
|
|
|
for (i = 0; i < cycle_gap; ++i) {
|
|
|
|
data_block_count += desc->data_blocks;
|
|
|
|
desc = prev_packet_desc(s, desc);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
end:
|
|
|
|
return data_block_count * s->pcm_frame_multiplier;
|
|
|
|
}
|
|
|
|
|
2019-07-22 03:37:02 +00:00
|
|
|
static void process_ctx_payloads(struct amdtp_stream *s,
|
2023-01-10 13:49:31 +00:00
|
|
|
const struct pkt_desc *desc,
|
2023-01-09 02:17:37 +00:00
|
|
|
unsigned int count)
|
2011-03-15 06:53:21 +00:00
|
|
|
{
|
2019-07-22 03:37:09 +00:00
|
|
|
struct snd_pcm_substream *pcm;
|
2023-01-10 13:49:31 +00:00
|
|
|
int i;
|
2019-07-22 03:37:01 +00:00
|
|
|
|
2019-07-22 03:37:09 +00:00
|
|
|
pcm = READ_ONCE(s->pcm);
|
2023-01-10 13:49:32 +00:00
|
|
|
s->process_ctx_payloads(s, desc, count, pcm);
|
2023-01-10 13:49:31 +00:00
|
|
|
|
|
|
|
if (pcm) {
|
|
|
|
unsigned int data_block_count = 0;
|
|
|
|
|
2023-01-10 13:49:33 +00:00
|
|
|
pcm->runtime->delay = compute_pcm_extra_delay(s, desc, count);
|
|
|
|
|
2023-01-10 13:49:31 +00:00
|
|
|
for (i = 0; i < count; ++i) {
|
|
|
|
data_block_count += desc->data_blocks;
|
|
|
|
desc = amdtp_stream_next_packet_desc(s, desc);
|
|
|
|
}
|
|
|
|
|
|
|
|
update_pcm_pointers(s, pcm, data_block_count * s->pcm_frame_multiplier);
|
|
|
|
}
|
2019-07-22 03:37:02 +00:00
|
|
|
}
|
|
|
|
|
2021-05-20 04:01:53 +00:00
|
|
|
static void process_rx_packets(struct fw_iso_context *context, u32 tstamp, size_t header_length,
|
|
|
|
void *header, void *private_data)
|
2019-07-22 03:37:02 +00:00
|
|
|
{
|
|
|
|
struct amdtp_stream *s = private_data;
|
2020-05-08 04:36:35 +00:00
|
|
|
const struct amdtp_domain *d = s->domain;
|
2019-07-22 03:37:02 +00:00
|
|
|
const __be32 *ctx_header = header;
|
2021-05-20 04:01:53 +00:00
|
|
|
const unsigned int events_per_period = d->events_per_period;
|
2019-10-18 06:19:10 +00:00
|
|
|
unsigned int event_count = s->ctx_data.rx.event_count;
|
2023-01-09 02:17:38 +00:00
|
|
|
struct pkt_desc *desc = s->packet_descs_cursor;
|
2021-05-20 04:01:49 +00:00
|
|
|
unsigned int pkt_header_length;
|
2019-10-17 15:54:13 +00:00
|
|
|
unsigned int packets;
|
2023-01-09 21:32:31 +00:00
|
|
|
u32 curr_cycle_time;
|
2021-05-27 12:32:53 +00:00
|
|
|
bool need_hw_irq;
|
2019-07-22 03:37:02 +00:00
|
|
|
int i;
|
|
|
|
|
|
|
|
if (s->packet_index < 0)
|
|
|
|
return;
|
|
|
|
|
2019-10-17 15:54:13 +00:00
|
|
|
// Calculate the number of packets in buffer and check XRUN.
|
|
|
|
packets = header_length / sizeof(*ctx_header);
|
|
|
|
|
2023-01-09 02:17:36 +00:00
|
|
|
generate_rx_packet_descs(s, desc, ctx_header, packets);
|
2019-07-22 03:37:02 +00:00
|
|
|
|
2023-01-09 02:17:36 +00:00
|
|
|
process_ctx_payloads(s, desc, packets);
|
2019-07-22 03:37:01 +00:00
|
|
|
|
2021-05-20 04:01:49 +00:00
|
|
|
if (!(s->flags & CIP_NO_HEADER))
|
|
|
|
pkt_header_length = IT_PKT_HEADER_SIZE_CIP;
|
|
|
|
else
|
|
|
|
pkt_header_length = 0;
|
|
|
|
|
2021-05-27 12:32:53 +00:00
|
|
|
if (s == d->irq_target) {
|
|
|
|
// At NO_PERIOD_WAKEUP mode, the packets for all IT/IR contexts are processed by
|
|
|
|
// the tasks of user process operating ALSA PCM character device by calling ioctl(2)
|
|
|
|
// with some requests, instead of scheduled hardware IRQ of an IT context.
|
|
|
|
struct snd_pcm_substream *pcm = READ_ONCE(s->pcm);
|
|
|
|
need_hw_irq = !pcm || !pcm->runtime->no_period_wakeup;
|
|
|
|
} else {
|
|
|
|
need_hw_irq = false;
|
|
|
|
}
|
|
|
|
|
2023-01-09 21:32:31 +00:00
|
|
|
if (trace_amdtp_packet_enabled())
|
|
|
|
(void)fw_card_read_cycle_time(fw_parent_device(s->unit)->card, &curr_cycle_time);
|
|
|
|
|
2019-07-22 03:37:01 +00:00
|
|
|
for (i = 0; i < packets; ++i) {
|
2024-07-25 15:56:40 +00:00
|
|
|
DEFINE_RAW_FLEX(struct fw_iso_packet, template, header, CIP_HEADER_QUADLETS);
|
2019-10-17 15:54:22 +00:00
|
|
|
bool sched_irq = false;
|
2011-03-15 06:53:21 +00:00
|
|
|
|
2024-03-26 01:59:34 +00:00
|
|
|
build_it_pkt_header(s, desc->cycle, template, pkt_header_length,
|
2019-07-22 03:36:59 +00:00
|
|
|
desc->data_blocks, desc->data_block_counter,
|
2023-01-09 21:32:31 +00:00
|
|
|
desc->syt, i, curr_cycle_time);
|
2019-05-24 09:03:41 +00:00
|
|
|
|
2020-05-08 04:36:28 +00:00
|
|
|
if (s == s->domain->irq_target) {
|
2019-10-18 06:19:10 +00:00
|
|
|
event_count += desc->data_blocks;
|
|
|
|
if (event_count >= events_per_period) {
|
|
|
|
event_count -= events_per_period;
|
2021-05-27 12:32:53 +00:00
|
|
|
sched_irq = need_hw_irq;
|
2019-10-18 06:19:10 +00:00
|
|
|
}
|
2019-10-17 15:54:22 +00:00
|
|
|
}
|
|
|
|
|
2024-03-26 01:59:34 +00:00
|
|
|
if (queue_out_packet(s, template, sched_irq) < 0) {
|
2019-05-21 14:57:35 +00:00
|
|
|
cancel_stream(s);
|
2015-05-22 14:00:53 +00:00
|
|
|
return;
|
|
|
|
}
|
2023-01-09 02:17:36 +00:00
|
|
|
|
|
|
|
desc = amdtp_stream_next_packet_desc(s, desc);
|
2014-04-25 13:44:48 +00:00
|
|
|
}
|
2015-05-22 14:00:53 +00:00
|
|
|
|
2019-10-18 06:19:10 +00:00
|
|
|
s->ctx_data.rx.event_count = event_count;
|
2023-01-09 02:17:38 +00:00
|
|
|
s->packet_descs_cursor = desc;
|
2011-03-15 06:53:21 +00:00
|
|
|
}
|
|
|
|
|
2021-05-20 04:01:53 +00:00
|
|
|
static void skip_rx_packets(struct fw_iso_context *context, u32 tstamp, size_t header_length,
|
|
|
|
void *header, void *private_data)
|
|
|
|
{
|
|
|
|
struct amdtp_stream *s = private_data;
|
|
|
|
struct amdtp_domain *d = s->domain;
|
|
|
|
const __be32 *ctx_header = header;
|
|
|
|
unsigned int packets;
|
|
|
|
unsigned int cycle;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
if (s->packet_index < 0)
|
|
|
|
return;
|
|
|
|
|
|
|
|
packets = header_length / sizeof(*ctx_header);
|
|
|
|
|
|
|
|
cycle = compute_ohci_it_cycle(ctx_header[packets - 1], s->queue_size);
|
|
|
|
s->next_cycle = increment_ohci_cycle_count(cycle, 1);
|
|
|
|
|
|
|
|
for (i = 0; i < packets; ++i) {
|
|
|
|
struct fw_iso_packet params = {
|
|
|
|
.header_length = 0,
|
|
|
|
.payload_length = 0,
|
|
|
|
};
|
|
|
|
bool sched_irq = (s == d->irq_target && i == packets - 1);
|
|
|
|
|
|
|
|
if (queue_out_packet(s, ¶ms, sched_irq) < 0) {
|
|
|
|
cancel_stream(s);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void irq_target_callback(struct fw_iso_context *context, u32 tstamp, size_t header_length,
|
|
|
|
void *header, void *private_data);
|
|
|
|
|
|
|
|
static void process_rx_packets_intermediately(struct fw_iso_context *context, u32 tstamp,
|
|
|
|
size_t header_length, void *header, void *private_data)
|
|
|
|
{
|
|
|
|
struct amdtp_stream *s = private_data;
|
|
|
|
struct amdtp_domain *d = s->domain;
|
|
|
|
__be32 *ctx_header = header;
|
|
|
|
const unsigned int queue_size = s->queue_size;
|
|
|
|
unsigned int packets;
|
|
|
|
unsigned int offset;
|
|
|
|
|
|
|
|
if (s->packet_index < 0)
|
|
|
|
return;
|
|
|
|
|
|
|
|
packets = header_length / sizeof(*ctx_header);
|
|
|
|
|
|
|
|
offset = 0;
|
|
|
|
while (offset < packets) {
|
|
|
|
unsigned int cycle = compute_ohci_it_cycle(ctx_header[offset], queue_size);
|
|
|
|
|
|
|
|
if (compare_ohci_cycle_count(cycle, d->processing_cycle.rx_start) >= 0)
|
|
|
|
break;
|
|
|
|
|
|
|
|
++offset;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (offset > 0) {
|
|
|
|
unsigned int length = sizeof(*ctx_header) * offset;
|
|
|
|
|
|
|
|
skip_rx_packets(context, tstamp, length, ctx_header, private_data);
|
|
|
|
if (amdtp_streaming_error(s))
|
|
|
|
return;
|
|
|
|
|
|
|
|
ctx_header += offset;
|
|
|
|
header_length -= length;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (offset < packets) {
|
2021-05-20 04:01:54 +00:00
|
|
|
s->ready_processing = true;
|
|
|
|
wake_up(&s->ready_wait);
|
|
|
|
|
2023-01-07 02:32:14 +00:00
|
|
|
if (d->replay.enable)
|
|
|
|
s->ctx_data.rx.cache_pos = 0;
|
|
|
|
|
2021-05-20 04:01:53 +00:00
|
|
|
process_rx_packets(context, tstamp, header_length, ctx_header, private_data);
|
|
|
|
if (amdtp_streaming_error(s))
|
|
|
|
return;
|
|
|
|
|
|
|
|
if (s == d->irq_target)
|
|
|
|
s->context->callback.sc = irq_target_callback;
|
|
|
|
else
|
|
|
|
s->context->callback.sc = process_rx_packets;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-05-20 04:01:50 +00:00
|
|
|
static void process_tx_packets(struct fw_iso_context *context, u32 tstamp, size_t header_length,
|
|
|
|
void *header, void *private_data)
|
2014-04-25 13:44:46 +00:00
|
|
|
{
|
|
|
|
struct amdtp_stream *s = private_data;
|
2019-03-17 11:25:06 +00:00
|
|
|
__be32 *ctx_header = header;
|
2023-01-09 02:17:38 +00:00
|
|
|
struct pkt_desc *desc = s->packet_descs_cursor;
|
2023-01-07 02:32:13 +00:00
|
|
|
unsigned int packet_count;
|
2021-05-18 13:00:47 +00:00
|
|
|
unsigned int desc_count;
|
2019-07-22 03:37:00 +00:00
|
|
|
int i;
|
|
|
|
int err;
|
2014-04-25 13:44:46 +00:00
|
|
|
|
2015-05-22 14:00:53 +00:00
|
|
|
if (s->packet_index < 0)
|
|
|
|
return;
|
|
|
|
|
2019-10-17 15:54:13 +00:00
|
|
|
// Calculate the number of packets in buffer and check XRUN.
|
2023-01-07 02:32:13 +00:00
|
|
|
packet_count = header_length / s->ctx_data.tx.ctx_header_size;
|
2016-05-09 12:12:45 +00:00
|
|
|
|
2021-05-18 13:00:47 +00:00
|
|
|
desc_count = 0;
|
2023-01-09 02:17:36 +00:00
|
|
|
err = generate_tx_packet_descs(s, desc, ctx_header, packet_count, &desc_count);
|
2019-07-22 03:37:00 +00:00
|
|
|
if (err < 0) {
|
|
|
|
if (err != -EAGAIN) {
|
|
|
|
cancel_stream(s);
|
|
|
|
return;
|
|
|
|
}
|
2019-07-22 03:37:01 +00:00
|
|
|
} else {
|
2021-05-27 12:26:09 +00:00
|
|
|
struct amdtp_domain *d = s->domain;
|
|
|
|
|
2023-01-09 02:17:36 +00:00
|
|
|
process_ctx_payloads(s, desc, desc_count);
|
2021-05-27 12:26:09 +00:00
|
|
|
|
|
|
|
if (d->replay.enable)
|
2023-01-09 02:17:36 +00:00
|
|
|
cache_seq(s, desc, desc_count);
|
2023-01-09 02:17:38 +00:00
|
|
|
|
|
|
|
for (i = 0; i < desc_count; ++i)
|
|
|
|
desc = amdtp_stream_next_packet_desc(s, desc);
|
|
|
|
s->packet_descs_cursor = desc;
|
2019-07-22 03:37:01 +00:00
|
|
|
}
|
|
|
|
|
2023-01-07 02:32:13 +00:00
|
|
|
for (i = 0; i < packet_count; ++i) {
|
2019-07-22 03:37:01 +00:00
|
|
|
struct fw_iso_packet params = {0};
|
2014-04-25 13:44:46 +00:00
|
|
|
|
2019-10-18 06:19:10 +00:00
|
|
|
if (queue_in_packet(s, ¶ms) < 0) {
|
2019-07-22 03:37:00 +00:00
|
|
|
cancel_stream(s);
|
|
|
|
return;
|
|
|
|
}
|
2014-04-25 13:44:49 +00:00
|
|
|
}
|
2019-10-18 06:19:10 +00:00
|
|
|
}
|
|
|
|
|
2021-05-20 04:01:50 +00:00
|
|
|
static void drop_tx_packets(struct fw_iso_context *context, u32 tstamp, size_t header_length,
|
|
|
|
void *header, void *private_data)
|
|
|
|
{
|
|
|
|
struct amdtp_stream *s = private_data;
|
|
|
|
const __be32 *ctx_header = header;
|
|
|
|
unsigned int packets;
|
|
|
|
unsigned int cycle;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
if (s->packet_index < 0)
|
|
|
|
return;
|
|
|
|
|
|
|
|
packets = header_length / s->ctx_data.tx.ctx_header_size;
|
|
|
|
|
|
|
|
ctx_header += (packets - 1) * s->ctx_data.tx.ctx_header_size / sizeof(*ctx_header);
|
|
|
|
cycle = compute_ohci_cycle_count(ctx_header[1]);
|
|
|
|
s->next_cycle = increment_ohci_cycle_count(cycle, 1);
|
|
|
|
|
|
|
|
for (i = 0; i < packets; ++i) {
|
|
|
|
struct fw_iso_packet params = {0};
|
|
|
|
|
|
|
|
if (queue_in_packet(s, ¶ms) < 0) {
|
|
|
|
cancel_stream(s);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void process_tx_packets_intermediately(struct fw_iso_context *context, u32 tstamp,
|
|
|
|
size_t header_length, void *header, void *private_data)
|
|
|
|
{
|
|
|
|
struct amdtp_stream *s = private_data;
|
|
|
|
struct amdtp_domain *d = s->domain;
|
|
|
|
__be32 *ctx_header;
|
|
|
|
unsigned int packets;
|
|
|
|
unsigned int offset;
|
|
|
|
|
|
|
|
if (s->packet_index < 0)
|
|
|
|
return;
|
|
|
|
|
|
|
|
packets = header_length / s->ctx_data.tx.ctx_header_size;
|
|
|
|
|
|
|
|
offset = 0;
|
|
|
|
ctx_header = header;
|
|
|
|
while (offset < packets) {
|
|
|
|
unsigned int cycle = compute_ohci_cycle_count(ctx_header[1]);
|
|
|
|
|
|
|
|
if (compare_ohci_cycle_count(cycle, d->processing_cycle.tx_start) >= 0)
|
|
|
|
break;
|
|
|
|
|
|
|
|
ctx_header += s->ctx_data.tx.ctx_header_size / sizeof(__be32);
|
|
|
|
++offset;
|
|
|
|
}
|
|
|
|
|
|
|
|
ctx_header = header;
|
|
|
|
|
|
|
|
if (offset > 0) {
|
|
|
|
size_t length = s->ctx_data.tx.ctx_header_size * offset;
|
|
|
|
|
|
|
|
drop_tx_packets(context, tstamp, length, ctx_header, s);
|
|
|
|
if (amdtp_streaming_error(s))
|
|
|
|
return;
|
|
|
|
|
|
|
|
ctx_header += length / sizeof(*ctx_header);
|
|
|
|
header_length -= length;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (offset < packets) {
|
2021-05-20 04:01:54 +00:00
|
|
|
s->ready_processing = true;
|
|
|
|
wake_up(&s->ready_wait);
|
|
|
|
|
2021-05-20 04:01:50 +00:00
|
|
|
process_tx_packets(context, tstamp, header_length, ctx_header, s);
|
|
|
|
if (amdtp_streaming_error(s))
|
|
|
|
return;
|
|
|
|
|
|
|
|
context->callback.sc = process_tx_packets;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-05-24 03:13:43 +00:00
|
|
|
static void drop_tx_packets_initially(struct fw_iso_context *context, u32 tstamp,
|
|
|
|
size_t header_length, void *header, void *private_data)
|
|
|
|
{
|
|
|
|
struct amdtp_stream *s = private_data;
|
|
|
|
struct amdtp_domain *d = s->domain;
|
|
|
|
__be32 *ctx_header;
|
|
|
|
unsigned int count;
|
|
|
|
unsigned int events;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
if (s->packet_index < 0)
|
|
|
|
return;
|
|
|
|
|
|
|
|
count = header_length / s->ctx_data.tx.ctx_header_size;
|
|
|
|
|
|
|
|
// Attempt to detect any event in the batch of packets.
|
|
|
|
events = 0;
|
|
|
|
ctx_header = header;
|
|
|
|
for (i = 0; i < count; ++i) {
|
|
|
|
unsigned int payload_quads =
|
|
|
|
(be32_to_cpu(*ctx_header) >> ISO_DATA_LENGTH_SHIFT) / sizeof(__be32);
|
|
|
|
unsigned int data_blocks;
|
|
|
|
|
|
|
|
if (s->flags & CIP_NO_HEADER) {
|
|
|
|
data_blocks = payload_quads / s->data_block_quadlets;
|
|
|
|
} else {
|
|
|
|
__be32 *cip_headers = ctx_header + IR_CTX_HEADER_DEFAULT_QUADLETS;
|
|
|
|
|
|
|
|
if (payload_quads < CIP_HEADER_QUADLETS) {
|
|
|
|
data_blocks = 0;
|
|
|
|
} else {
|
|
|
|
payload_quads -= CIP_HEADER_QUADLETS;
|
|
|
|
|
|
|
|
if (s->flags & CIP_UNAWARE_SYT) {
|
|
|
|
data_blocks = payload_quads / s->data_block_quadlets;
|
|
|
|
} else {
|
|
|
|
u32 cip1 = be32_to_cpu(cip_headers[1]);
|
|
|
|
|
|
|
|
// NODATA packet can includes any data blocks but they are
|
|
|
|
// not available as event.
|
|
|
|
if ((cip1 & CIP_NO_DATA) == CIP_NO_DATA)
|
|
|
|
data_blocks = 0;
|
|
|
|
else
|
|
|
|
data_blocks = payload_quads / s->data_block_quadlets;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
events += data_blocks;
|
|
|
|
|
|
|
|
ctx_header += s->ctx_data.tx.ctx_header_size / sizeof(__be32);
|
|
|
|
}
|
|
|
|
|
|
|
|
drop_tx_packets(context, tstamp, header_length, header, s);
|
|
|
|
|
|
|
|
if (events > 0)
|
|
|
|
s->ctx_data.tx.event_starts = true;
|
|
|
|
|
|
|
|
// Decide the cycle count to begin processing content of packet in IR contexts.
|
|
|
|
{
|
|
|
|
unsigned int stream_count = 0;
|
|
|
|
unsigned int event_starts_count = 0;
|
|
|
|
unsigned int cycle = UINT_MAX;
|
|
|
|
|
|
|
|
list_for_each_entry(s, &d->streams, list) {
|
|
|
|
if (s->direction == AMDTP_IN_STREAM) {
|
|
|
|
++stream_count;
|
|
|
|
if (s->ctx_data.tx.event_starts)
|
|
|
|
++event_starts_count;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (stream_count == event_starts_count) {
|
|
|
|
unsigned int next_cycle;
|
|
|
|
|
|
|
|
list_for_each_entry(s, &d->streams, list) {
|
|
|
|
if (s->direction != AMDTP_IN_STREAM)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
next_cycle = increment_ohci_cycle_count(s->next_cycle,
|
|
|
|
d->processing_cycle.tx_init_skip);
|
|
|
|
if (cycle == UINT_MAX ||
|
|
|
|
compare_ohci_cycle_count(next_cycle, cycle) > 0)
|
|
|
|
cycle = next_cycle;
|
|
|
|
|
|
|
|
s->context->callback.sc = process_tx_packets_intermediately;
|
|
|
|
}
|
|
|
|
|
|
|
|
d->processing_cycle.tx_start = cycle;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-05-20 04:01:53 +00:00
|
|
|
static void process_ctxs_in_domain(struct amdtp_domain *d)
|
2019-10-18 06:19:10 +00:00
|
|
|
{
|
|
|
|
struct amdtp_stream *s;
|
|
|
|
|
|
|
|
list_for_each_entry(s, &d->streams, list) {
|
2021-05-20 04:01:53 +00:00
|
|
|
if (s != d->irq_target && amdtp_stream_running(s))
|
2019-10-18 06:19:10 +00:00
|
|
|
fw_iso_context_flush_completions(s->context);
|
2021-05-20 04:01:53 +00:00
|
|
|
|
|
|
|
if (amdtp_streaming_error(s))
|
|
|
|
goto error;
|
2019-10-18 06:19:10 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
return;
|
|
|
|
error:
|
2021-05-20 04:01:53 +00:00
|
|
|
if (amdtp_stream_running(d->irq_target))
|
|
|
|
cancel_stream(d->irq_target);
|
2019-10-18 06:19:10 +00:00
|
|
|
|
|
|
|
list_for_each_entry(s, &d->streams, list) {
|
|
|
|
if (amdtp_stream_running(s))
|
|
|
|
cancel_stream(s);
|
|
|
|
}
|
2014-04-25 13:44:46 +00:00
|
|
|
}
|
|
|
|
|
2021-05-20 04:01:53 +00:00
|
|
|
static void irq_target_callback(struct fw_iso_context *context, u32 tstamp, size_t header_length,
|
|
|
|
void *header, void *private_data)
|
|
|
|
{
|
|
|
|
struct amdtp_stream *s = private_data;
|
|
|
|
struct amdtp_domain *d = s->domain;
|
|
|
|
|
|
|
|
process_rx_packets(context, tstamp, header_length, header, private_data);
|
|
|
|
process_ctxs_in_domain(d);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void irq_target_callback_intermediately(struct fw_iso_context *context, u32 tstamp,
|
|
|
|
size_t header_length, void *header, void *private_data)
|
|
|
|
{
|
|
|
|
struct amdtp_stream *s = private_data;
|
|
|
|
struct amdtp_domain *d = s->domain;
|
|
|
|
|
|
|
|
process_rx_packets_intermediately(context, tstamp, header_length, header, private_data);
|
|
|
|
process_ctxs_in_domain(d);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void irq_target_callback_skip(struct fw_iso_context *context, u32 tstamp,
|
|
|
|
size_t header_length, void *header, void *private_data)
|
|
|
|
{
|
|
|
|
struct amdtp_stream *s = private_data;
|
|
|
|
struct amdtp_domain *d = s->domain;
|
ALSA: firewire-lib: replay sequence of incoming packets for outgoing packets
ALSA IEC 61883-1/6 packet streaming engine uses pre-computed parameters
ideal for nominal sampling transfer frequency (STF) to transfer packets
to device since it was added 2011. As a result of user experience for a
decade, it is clear that the sequence is not suitable to some actual
devices. It takes the devices to generate noise, and causes any type of
discontinuity in the series of packet transferred from the device. It's
required for the engine to transfer packets according to effective STF.
The effective STF is given by media clock recovered by the sequence of
packet transferred from the target device. In the previous commit, the
sequence is already cached. The media clock recovery can be achieved by
analyzing the sequence.
In technological world, many ideas are proposed for media clock recovery.
However, the small part of them could be actually adopted in our case
since floating point arithmetic is not mostly available in Linux kernel
land.
This commit adopts the simple way from them; sequence replay, which means
that the sequence of parameters from incoming packet is used as is to
transfer outgoing packets. The media clock is not computed internally,
but the sequence of outgoing packet superficially looks to be generated by
the media clock.
The association between source and destination is decided when starting
AMDTP domain. When the target device supports a pair of isochronous packet
streams, the tx stream is source and the rx stream is destination. When it
supports two pair of streams, each of tx stream is associated to
corresponding rx stream in its order. When it supports less number of tx
streams than rx streams, the fist tx stream is selected for all of rx
streams. When it supports more tx streams than rx streams, the first tx
packet is associated to the rx stream.
As I noted in previous commit, the sequence of parameters from incoming
packet is different between devices, time to time. It is worse idea to
replay the sequence of parameters from a device for the sequence of
packet to the other devices even if they are in the same category of
device.
Signed-off-by: Takashi Sakamoto <o-takashi@sakamocchi.jp>
Link: https://lore.kernel.org/r/20210527122611.173711-3-o-takashi@sakamocchi.jp
Signed-off-by: Takashi Iwai <tiwai@suse.de>
2021-05-27 12:26:10 +00:00
|
|
|
bool ready_to_start;
|
2021-05-20 04:01:53 +00:00
|
|
|
|
|
|
|
skip_rx_packets(context, tstamp, header_length, header, private_data);
|
|
|
|
process_ctxs_in_domain(d);
|
|
|
|
|
2021-05-27 12:26:11 +00:00
|
|
|
if (d->replay.enable && !d->replay.on_the_fly) {
|
ALSA: firewire-lib: replay sequence of incoming packets for outgoing packets
ALSA IEC 61883-1/6 packet streaming engine uses pre-computed parameters
ideal for nominal sampling transfer frequency (STF) to transfer packets
to device since it was added 2011. As a result of user experience for a
decade, it is clear that the sequence is not suitable to some actual
devices. It takes the devices to generate noise, and causes any type of
discontinuity in the series of packet transferred from the device. It's
required for the engine to transfer packets according to effective STF.
The effective STF is given by media clock recovered by the sequence of
packet transferred from the target device. In the previous commit, the
sequence is already cached. The media clock recovery can be achieved by
analyzing the sequence.
In technological world, many ideas are proposed for media clock recovery.
However, the small part of them could be actually adopted in our case
since floating point arithmetic is not mostly available in Linux kernel
land.
This commit adopts the simple way from them; sequence replay, which means
that the sequence of parameters from incoming packet is used as is to
transfer outgoing packets. The media clock is not computed internally,
but the sequence of outgoing packet superficially looks to be generated by
the media clock.
The association between source and destination is decided when starting
AMDTP domain. When the target device supports a pair of isochronous packet
streams, the tx stream is source and the rx stream is destination. When it
supports two pair of streams, each of tx stream is associated to
corresponding rx stream in its order. When it supports less number of tx
streams than rx streams, the fist tx stream is selected for all of rx
streams. When it supports more tx streams than rx streams, the first tx
packet is associated to the rx stream.
As I noted in previous commit, the sequence of parameters from incoming
packet is different between devices, time to time. It is worse idea to
replay the sequence of parameters from a device for the sequence of
packet to the other devices even if they are in the same category of
device.
Signed-off-by: Takashi Sakamoto <o-takashi@sakamocchi.jp>
Link: https://lore.kernel.org/r/20210527122611.173711-3-o-takashi@sakamocchi.jp
Signed-off-by: Takashi Iwai <tiwai@suse.de>
2021-05-27 12:26:10 +00:00
|
|
|
unsigned int rx_count = 0;
|
|
|
|
unsigned int rx_ready_count = 0;
|
|
|
|
struct amdtp_stream *rx;
|
|
|
|
|
|
|
|
list_for_each_entry(rx, &d->streams, list) {
|
|
|
|
struct amdtp_stream *tx;
|
|
|
|
unsigned int cached_cycles;
|
|
|
|
|
|
|
|
if (rx->direction != AMDTP_OUT_STREAM)
|
|
|
|
continue;
|
|
|
|
++rx_count;
|
|
|
|
|
|
|
|
tx = rx->ctx_data.rx.replay_target;
|
|
|
|
cached_cycles = calculate_cached_cycle_count(tx, 0);
|
|
|
|
if (cached_cycles > tx->ctx_data.tx.cache.size / 2)
|
|
|
|
++rx_ready_count;
|
|
|
|
}
|
|
|
|
|
|
|
|
ready_to_start = (rx_count == rx_ready_count);
|
|
|
|
} else {
|
|
|
|
ready_to_start = true;
|
|
|
|
}
|
|
|
|
|
2021-05-20 04:01:53 +00:00
|
|
|
// Decide the cycle count to begin processing content of packet in IT contexts. All of IT
|
|
|
|
// contexts are expected to start and get callback when reaching here.
|
ALSA: firewire-lib: replay sequence of incoming packets for outgoing packets
ALSA IEC 61883-1/6 packet streaming engine uses pre-computed parameters
ideal for nominal sampling transfer frequency (STF) to transfer packets
to device since it was added 2011. As a result of user experience for a
decade, it is clear that the sequence is not suitable to some actual
devices. It takes the devices to generate noise, and causes any type of
discontinuity in the series of packet transferred from the device. It's
required for the engine to transfer packets according to effective STF.
The effective STF is given by media clock recovered by the sequence of
packet transferred from the target device. In the previous commit, the
sequence is already cached. The media clock recovery can be achieved by
analyzing the sequence.
In technological world, many ideas are proposed for media clock recovery.
However, the small part of them could be actually adopted in our case
since floating point arithmetic is not mostly available in Linux kernel
land.
This commit adopts the simple way from them; sequence replay, which means
that the sequence of parameters from incoming packet is used as is to
transfer outgoing packets. The media clock is not computed internally,
but the sequence of outgoing packet superficially looks to be generated by
the media clock.
The association between source and destination is decided when starting
AMDTP domain. When the target device supports a pair of isochronous packet
streams, the tx stream is source and the rx stream is destination. When it
supports two pair of streams, each of tx stream is associated to
corresponding rx stream in its order. When it supports less number of tx
streams than rx streams, the fist tx stream is selected for all of rx
streams. When it supports more tx streams than rx streams, the first tx
packet is associated to the rx stream.
As I noted in previous commit, the sequence of parameters from incoming
packet is different between devices, time to time. It is worse idea to
replay the sequence of parameters from a device for the sequence of
packet to the other devices even if they are in the same category of
device.
Signed-off-by: Takashi Sakamoto <o-takashi@sakamocchi.jp>
Link: https://lore.kernel.org/r/20210527122611.173711-3-o-takashi@sakamocchi.jp
Signed-off-by: Takashi Iwai <tiwai@suse.de>
2021-05-27 12:26:10 +00:00
|
|
|
if (ready_to_start) {
|
|
|
|
unsigned int cycle = s->next_cycle;
|
|
|
|
list_for_each_entry(s, &d->streams, list) {
|
|
|
|
if (s->direction != AMDTP_OUT_STREAM)
|
|
|
|
continue;
|
2021-05-20 04:01:53 +00:00
|
|
|
|
ALSA: firewire-lib: replay sequence of incoming packets for outgoing packets
ALSA IEC 61883-1/6 packet streaming engine uses pre-computed parameters
ideal for nominal sampling transfer frequency (STF) to transfer packets
to device since it was added 2011. As a result of user experience for a
decade, it is clear that the sequence is not suitable to some actual
devices. It takes the devices to generate noise, and causes any type of
discontinuity in the series of packet transferred from the device. It's
required for the engine to transfer packets according to effective STF.
The effective STF is given by media clock recovered by the sequence of
packet transferred from the target device. In the previous commit, the
sequence is already cached. The media clock recovery can be achieved by
analyzing the sequence.
In technological world, many ideas are proposed for media clock recovery.
However, the small part of them could be actually adopted in our case
since floating point arithmetic is not mostly available in Linux kernel
land.
This commit adopts the simple way from them; sequence replay, which means
that the sequence of parameters from incoming packet is used as is to
transfer outgoing packets. The media clock is not computed internally,
but the sequence of outgoing packet superficially looks to be generated by
the media clock.
The association between source and destination is decided when starting
AMDTP domain. When the target device supports a pair of isochronous packet
streams, the tx stream is source and the rx stream is destination. When it
supports two pair of streams, each of tx stream is associated to
corresponding rx stream in its order. When it supports less number of tx
streams than rx streams, the fist tx stream is selected for all of rx
streams. When it supports more tx streams than rx streams, the first tx
packet is associated to the rx stream.
As I noted in previous commit, the sequence of parameters from incoming
packet is different between devices, time to time. It is worse idea to
replay the sequence of parameters from a device for the sequence of
packet to the other devices even if they are in the same category of
device.
Signed-off-by: Takashi Sakamoto <o-takashi@sakamocchi.jp>
Link: https://lore.kernel.org/r/20210527122611.173711-3-o-takashi@sakamocchi.jp
Signed-off-by: Takashi Iwai <tiwai@suse.de>
2021-05-27 12:26:10 +00:00
|
|
|
if (compare_ohci_cycle_count(s->next_cycle, cycle) > 0)
|
|
|
|
cycle = s->next_cycle;
|
2021-05-20 04:01:53 +00:00
|
|
|
|
ALSA: firewire-lib: replay sequence of incoming packets for outgoing packets
ALSA IEC 61883-1/6 packet streaming engine uses pre-computed parameters
ideal for nominal sampling transfer frequency (STF) to transfer packets
to device since it was added 2011. As a result of user experience for a
decade, it is clear that the sequence is not suitable to some actual
devices. It takes the devices to generate noise, and causes any type of
discontinuity in the series of packet transferred from the device. It's
required for the engine to transfer packets according to effective STF.
The effective STF is given by media clock recovered by the sequence of
packet transferred from the target device. In the previous commit, the
sequence is already cached. The media clock recovery can be achieved by
analyzing the sequence.
In technological world, many ideas are proposed for media clock recovery.
However, the small part of them could be actually adopted in our case
since floating point arithmetic is not mostly available in Linux kernel
land.
This commit adopts the simple way from them; sequence replay, which means
that the sequence of parameters from incoming packet is used as is to
transfer outgoing packets. The media clock is not computed internally,
but the sequence of outgoing packet superficially looks to be generated by
the media clock.
The association between source and destination is decided when starting
AMDTP domain. When the target device supports a pair of isochronous packet
streams, the tx stream is source and the rx stream is destination. When it
supports two pair of streams, each of tx stream is associated to
corresponding rx stream in its order. When it supports less number of tx
streams than rx streams, the fist tx stream is selected for all of rx
streams. When it supports more tx streams than rx streams, the first tx
packet is associated to the rx stream.
As I noted in previous commit, the sequence of parameters from incoming
packet is different between devices, time to time. It is worse idea to
replay the sequence of parameters from a device for the sequence of
packet to the other devices even if they are in the same category of
device.
Signed-off-by: Takashi Sakamoto <o-takashi@sakamocchi.jp>
Link: https://lore.kernel.org/r/20210527122611.173711-3-o-takashi@sakamocchi.jp
Signed-off-by: Takashi Iwai <tiwai@suse.de>
2021-05-27 12:26:10 +00:00
|
|
|
if (s == d->irq_target)
|
|
|
|
s->context->callback.sc = irq_target_callback_intermediately;
|
|
|
|
else
|
|
|
|
s->context->callback.sc = process_rx_packets_intermediately;
|
|
|
|
}
|
2021-05-20 04:01:53 +00:00
|
|
|
|
ALSA: firewire-lib: replay sequence of incoming packets for outgoing packets
ALSA IEC 61883-1/6 packet streaming engine uses pre-computed parameters
ideal for nominal sampling transfer frequency (STF) to transfer packets
to device since it was added 2011. As a result of user experience for a
decade, it is clear that the sequence is not suitable to some actual
devices. It takes the devices to generate noise, and causes any type of
discontinuity in the series of packet transferred from the device. It's
required for the engine to transfer packets according to effective STF.
The effective STF is given by media clock recovered by the sequence of
packet transferred from the target device. In the previous commit, the
sequence is already cached. The media clock recovery can be achieved by
analyzing the sequence.
In technological world, many ideas are proposed for media clock recovery.
However, the small part of them could be actually adopted in our case
since floating point arithmetic is not mostly available in Linux kernel
land.
This commit adopts the simple way from them; sequence replay, which means
that the sequence of parameters from incoming packet is used as is to
transfer outgoing packets. The media clock is not computed internally,
but the sequence of outgoing packet superficially looks to be generated by
the media clock.
The association between source and destination is decided when starting
AMDTP domain. When the target device supports a pair of isochronous packet
streams, the tx stream is source and the rx stream is destination. When it
supports two pair of streams, each of tx stream is associated to
corresponding rx stream in its order. When it supports less number of tx
streams than rx streams, the fist tx stream is selected for all of rx
streams. When it supports more tx streams than rx streams, the first tx
packet is associated to the rx stream.
As I noted in previous commit, the sequence of parameters from incoming
packet is different between devices, time to time. It is worse idea to
replay the sequence of parameters from a device for the sequence of
packet to the other devices even if they are in the same category of
device.
Signed-off-by: Takashi Sakamoto <o-takashi@sakamocchi.jp>
Link: https://lore.kernel.org/r/20210527122611.173711-3-o-takashi@sakamocchi.jp
Signed-off-by: Takashi Iwai <tiwai@suse.de>
2021-05-27 12:26:10 +00:00
|
|
|
d->processing_cycle.rx_start = cycle;
|
|
|
|
}
|
2021-05-20 04:01:53 +00:00
|
|
|
}
|
|
|
|
|
2021-05-24 03:13:44 +00:00
|
|
|
// This is executed one time. For in-stream, first packet has come. For out-stream, prepared to
|
|
|
|
// transmit first packet.
|
2014-04-25 13:44:49 +00:00
|
|
|
static void amdtp_stream_first_callback(struct fw_iso_context *context,
|
ALSA: firewire-lib: compute the value of second field in cycle count for IT context
In callback function of isochronous context, u32 variable is passed for
cycle count. The value of this variable comes from DMA descriptors of 1394
Open Host Controller Interface (1394 OHCI). In the specification, DMA
descriptors transport lower 3 bits for second field and full cycle field in
16 bits field, therefore 16 bits of the u32 variable are available. The
value for second is modulo 8, and the value for cycle is modulo 8,000.
Currently, ALSA firewire-lib module don't use the value of the second
field, because the value is useless to calculate presentation timestamp in
IEC 61883-6. However, the value may be useful for debugging. In later
commit, it will be printed with the other parameters for debugging.
This commit makes this module to handle the whole cycle count including
second. The value is calculated by cycle unit. The existed code is already
written with ignoring the value of second, thus this commit causes no
issues.
Signed-off-by: Takashi Sakamoto <o-takashi@sakamocchi.jp>
Signed-off-by: Takashi Iwai <tiwai@suse.de>
2016-05-09 12:12:44 +00:00
|
|
|
u32 tstamp, size_t header_length,
|
2014-04-25 13:44:49 +00:00
|
|
|
void *header, void *private_data)
|
|
|
|
{
|
|
|
|
struct amdtp_stream *s = private_data;
|
2021-05-20 04:01:50 +00:00
|
|
|
struct amdtp_domain *d = s->domain;
|
2014-04-25 13:44:49 +00:00
|
|
|
|
2017-03-22 12:30:15 +00:00
|
|
|
if (s->direction == AMDTP_IN_STREAM) {
|
2021-05-24 03:13:43 +00:00
|
|
|
context->callback.sc = drop_tx_packets_initially;
|
2017-03-22 12:30:15 +00:00
|
|
|
} else {
|
2021-05-20 04:01:50 +00:00
|
|
|
if (s == d->irq_target)
|
2021-05-20 04:01:53 +00:00
|
|
|
context->callback.sc = irq_target_callback_skip;
|
2020-05-08 04:36:28 +00:00
|
|
|
else
|
2021-05-20 04:01:53 +00:00
|
|
|
context->callback.sc = skip_rx_packets;
|
2017-03-22 12:30:15 +00:00
|
|
|
}
|
|
|
|
|
ALSA: firewire-lib: compute the value of second field in cycle count for IT context
In callback function of isochronous context, u32 variable is passed for
cycle count. The value of this variable comes from DMA descriptors of 1394
Open Host Controller Interface (1394 OHCI). In the specification, DMA
descriptors transport lower 3 bits for second field and full cycle field in
16 bits field, therefore 16 bits of the u32 variable are available. The
value for second is modulo 8, and the value for cycle is modulo 8,000.
Currently, ALSA firewire-lib module don't use the value of the second
field, because the value is useless to calculate presentation timestamp in
IEC 61883-6. However, the value may be useful for debugging. In later
commit, it will be printed with the other parameters for debugging.
This commit makes this module to handle the whole cycle count including
second. The value is calculated by cycle unit. The existed code is already
written with ignoring the value of second, thus this commit causes no
issues.
Signed-off-by: Takashi Sakamoto <o-takashi@sakamocchi.jp>
Signed-off-by: Takashi Iwai <tiwai@suse.de>
2016-05-09 12:12:44 +00:00
|
|
|
context->callback.sc(context, tstamp, header_length, header, s);
|
2014-04-25 13:44:49 +00:00
|
|
|
}
|
|
|
|
|
2011-03-15 06:53:21 +00:00
|
|
|
/**
|
2014-04-25 13:44:42 +00:00
|
|
|
* amdtp_stream_start - start transferring packets
|
|
|
|
* @s: the AMDTP stream to start
|
2011-03-15 06:53:21 +00:00
|
|
|
* @channel: the isochronous channel on the bus
|
|
|
|
* @speed: firewire speed code
|
2020-05-08 04:36:29 +00:00
|
|
|
* @queue_size: The number of packets in the queue.
|
|
|
|
* @idle_irq_interval: the interval to queue packet during initial state.
|
2011-03-15 06:53:21 +00:00
|
|
|
*
|
|
|
|
* The stream cannot be started until it has been configured with
|
2014-04-25 13:44:42 +00:00
|
|
|
* amdtp_stream_set_parameters() and it must be started before any PCM or MIDI
|
|
|
|
* device can be started.
|
2011-03-15 06:53:21 +00:00
|
|
|
*/
|
2019-10-17 15:54:13 +00:00
|
|
|
static int amdtp_stream_start(struct amdtp_stream *s, int channel, int speed,
|
2021-05-20 04:01:52 +00:00
|
|
|
unsigned int queue_size, unsigned int idle_irq_interval)
|
2011-03-15 06:53:21 +00:00
|
|
|
{
|
2020-05-08 04:36:28 +00:00
|
|
|
bool is_irq_target = (s == s->domain->irq_target);
|
2019-05-21 14:57:34 +00:00
|
|
|
unsigned int ctx_header_size;
|
2019-05-22 14:17:07 +00:00
|
|
|
unsigned int max_ctx_payload_size;
|
2014-04-25 13:44:46 +00:00
|
|
|
enum dma_data_direction dir;
|
2023-01-09 02:17:36 +00:00
|
|
|
struct pkt_desc *descs;
|
|
|
|
int i, type, tag, err;
|
2011-03-15 06:53:21 +00:00
|
|
|
|
|
|
|
mutex_lock(&s->mutex);
|
|
|
|
|
2014-04-25 13:44:42 +00:00
|
|
|
if (WARN_ON(amdtp_stream_running(s) ||
|
2014-04-25 13:44:45 +00:00
|
|
|
(s->data_block_quadlets < 1))) {
|
2011-03-15 06:53:21 +00:00
|
|
|
err = -EBADFD;
|
|
|
|
goto err_unlock;
|
|
|
|
}
|
|
|
|
|
2019-05-21 14:57:34 +00:00
|
|
|
if (s->direction == AMDTP_IN_STREAM) {
|
2019-10-18 06:19:10 +00:00
|
|
|
// NOTE: IT context should be used for constant IRQ.
|
|
|
|
if (is_irq_target) {
|
|
|
|
err = -EINVAL;
|
|
|
|
goto err_unlock;
|
|
|
|
}
|
|
|
|
|
2014-04-25 13:45:16 +00:00
|
|
|
s->data_block_counter = UINT_MAX;
|
2019-05-21 14:57:34 +00:00
|
|
|
} else {
|
2014-04-25 13:45:16 +00:00
|
|
|
s->data_block_counter = 0;
|
2019-05-21 14:57:34 +00:00
|
|
|
}
|
2011-03-15 06:53:21 +00:00
|
|
|
|
2021-05-13 12:56:51 +00:00
|
|
|
// initialize packet buffer.
|
2014-04-25 13:44:46 +00:00
|
|
|
if (s->direction == AMDTP_IN_STREAM) {
|
|
|
|
dir = DMA_FROM_DEVICE;
|
|
|
|
type = FW_ISO_CONTEXT_RECEIVE;
|
2021-05-20 04:01:48 +00:00
|
|
|
if (!(s->flags & CIP_NO_HEADER))
|
2019-05-22 14:17:07 +00:00
|
|
|
ctx_header_size = IR_CTX_HEADER_SIZE_CIP;
|
2021-05-20 04:01:48 +00:00
|
|
|
else
|
2019-05-22 14:17:07 +00:00
|
|
|
ctx_header_size = IR_CTX_HEADER_SIZE_NO_CIP;
|
2014-04-25 13:44:46 +00:00
|
|
|
} else {
|
|
|
|
dir = DMA_TO_DEVICE;
|
|
|
|
type = FW_ISO_CONTEXT_TRANSMIT;
|
2019-05-21 14:57:36 +00:00
|
|
|
ctx_header_size = 0; // No effect for IT context.
|
2019-05-23 15:14:40 +00:00
|
|
|
}
|
2021-05-20 04:01:48 +00:00
|
|
|
max_ctx_payload_size = amdtp_stream_get_max_ctx_payload_size(s);
|
2019-05-22 14:17:07 +00:00
|
|
|
|
2021-05-20 04:01:48 +00:00
|
|
|
err = iso_packets_buffer_init(&s->buffer, s->unit, queue_size, max_ctx_payload_size, dir);
|
2011-03-15 06:53:21 +00:00
|
|
|
if (err < 0)
|
|
|
|
goto err_unlock;
|
2020-05-08 04:36:29 +00:00
|
|
|
s->queue_size = queue_size;
|
2019-10-18 06:19:10 +00:00
|
|
|
|
2011-03-15 06:53:21 +00:00
|
|
|
s->context = fw_iso_context_create(fw_parent_device(s->unit)->card,
|
2019-05-21 14:57:34 +00:00
|
|
|
type, channel, speed, ctx_header_size,
|
2020-05-08 04:36:28 +00:00
|
|
|
amdtp_stream_first_callback, s);
|
2011-03-15 06:53:21 +00:00
|
|
|
if (IS_ERR(s->context)) {
|
|
|
|
err = PTR_ERR(s->context);
|
|
|
|
if (err == -EBUSY)
|
|
|
|
dev_err(&s->unit->device,
|
2014-04-25 13:44:42 +00:00
|
|
|
"no free stream on this controller\n");
|
2011-03-15 06:53:21 +00:00
|
|
|
goto err_buffer;
|
|
|
|
}
|
|
|
|
|
2014-04-25 13:44:42 +00:00
|
|
|
amdtp_stream_update(s);
|
2011-03-15 06:53:21 +00:00
|
|
|
|
2019-05-21 14:57:34 +00:00
|
|
|
if (s->direction == AMDTP_IN_STREAM) {
|
2019-05-22 14:17:07 +00:00
|
|
|
s->ctx_data.tx.max_ctx_payload_length = max_ctx_payload_size;
|
2019-05-21 14:57:34 +00:00
|
|
|
s->ctx_data.tx.ctx_header_size = ctx_header_size;
|
2021-05-24 03:13:43 +00:00
|
|
|
s->ctx_data.tx.event_starts = false;
|
2021-05-27 12:26:09 +00:00
|
|
|
|
|
|
|
if (s->domain->replay.enable) {
|
|
|
|
// struct fw_iso_context.drop_overflow_headers is false therefore it's
|
|
|
|
// possible to cache much unexpectedly.
|
|
|
|
s->ctx_data.tx.cache.size = max_t(unsigned int, s->syt_interval * 2,
|
|
|
|
queue_size * 3 / 2);
|
2023-01-07 02:32:13 +00:00
|
|
|
s->ctx_data.tx.cache.pos = 0;
|
2021-05-27 12:26:09 +00:00
|
|
|
s->ctx_data.tx.cache.descs = kcalloc(s->ctx_data.tx.cache.size,
|
|
|
|
sizeof(*s->ctx_data.tx.cache.descs), GFP_KERNEL);
|
2021-06-05 12:46:39 +00:00
|
|
|
if (!s->ctx_data.tx.cache.descs) {
|
|
|
|
err = -ENOMEM;
|
2021-05-27 12:26:09 +00:00
|
|
|
goto err_context;
|
2021-06-05 12:46:39 +00:00
|
|
|
}
|
2021-05-27 12:26:09 +00:00
|
|
|
}
|
2021-05-20 04:01:52 +00:00
|
|
|
} else {
|
2021-05-22 01:32:59 +00:00
|
|
|
static const struct {
|
|
|
|
unsigned int data_block;
|
|
|
|
unsigned int syt_offset;
|
|
|
|
} *entry, initial_state[] = {
|
|
|
|
[CIP_SFC_32000] = { 4, 3072 },
|
|
|
|
[CIP_SFC_48000] = { 6, 1024 },
|
|
|
|
[CIP_SFC_96000] = { 12, 1024 },
|
|
|
|
[CIP_SFC_192000] = { 24, 1024 },
|
|
|
|
[CIP_SFC_44100] = { 0, 67 },
|
|
|
|
[CIP_SFC_88200] = { 0, 67 },
|
|
|
|
[CIP_SFC_176400] = { 0, 67 },
|
|
|
|
};
|
|
|
|
|
|
|
|
s->ctx_data.rx.seq.descs = kcalloc(queue_size, sizeof(*s->ctx_data.rx.seq.descs), GFP_KERNEL);
|
2021-06-05 12:46:39 +00:00
|
|
|
if (!s->ctx_data.rx.seq.descs) {
|
|
|
|
err = -ENOMEM;
|
2021-05-22 01:32:59 +00:00
|
|
|
goto err_context;
|
2021-06-05 12:46:39 +00:00
|
|
|
}
|
2021-05-22 01:32:59 +00:00
|
|
|
s->ctx_data.rx.seq.size = queue_size;
|
2023-01-07 02:32:11 +00:00
|
|
|
s->ctx_data.rx.seq.pos = 0;
|
2021-05-22 01:32:59 +00:00
|
|
|
|
|
|
|
entry = &initial_state[s->sfc];
|
|
|
|
s->ctx_data.rx.data_block_state = entry->data_block;
|
|
|
|
s->ctx_data.rx.syt_offset_state = entry->syt_offset;
|
|
|
|
s->ctx_data.rx.last_syt_offset = TICKS_PER_CYCLE;
|
|
|
|
|
2021-05-20 04:01:52 +00:00
|
|
|
s->ctx_data.rx.event_count = 0;
|
2019-05-21 14:57:34 +00:00
|
|
|
}
|
2018-04-29 06:01:46 +00:00
|
|
|
|
ALSA: firewire-lib: add no-header packet processing
As long as investigating Fireface 400, IEC 61883-1/6 is not applied to
its packet streaming protocol. Remarks of the specific protocol are:
* Each packet doesn't include CIP headers.
* 64,0 and 128,0 kHz are supported.
* The device doesn't necessarily transmit 8,000 packets per second.
* 0, 1, 2, 3 are used as tag for rx isochronous packets, however 0 is
used for tx isochronous packets.
On the other hand, there's a common feature. The number of data blocks
transferred in a second is the same as sampling transmission frequency.
Current ALSA IEC 61883-1/6 engine already has a method to calculate it and
this driver can utilize it for rx packets, as well as tx packets.
This commit adds support for the transferring protocol. CIP_NO_HEADERS
flag is newly added. When this flag is set:
* Both of 0 (without CIP header) and 1 (with CIP header) are used as tag
to handle incoming isochronous packet.
* 0 (without CIP header) is used as tag to transfer outgoing isochronous
packet.
* Skip CIP header evaluation.
* Use unique way to calculate the quadlets of isochronous packet payload.
In ALSA PCM interface, 128.0 kHz is not supported, and the ALSA
IEC 61883-1/6 engine doesn't support 64.0 kHz. These modes are dropped.
The sequence of rx packet has a remarkable quirk about tag. This will be
described in later commits.
Signed-off-by: Takashi Sakamoto <o-takashi@sakamocchi.jp>
Signed-off-by: Takashi Iwai <tiwai@suse.de>
2017-03-31 13:06:07 +00:00
|
|
|
if (s->flags & CIP_NO_HEADER)
|
|
|
|
s->tag = TAG_NO_CIP_HEADER;
|
|
|
|
else
|
|
|
|
s->tag = TAG_CIP;
|
|
|
|
|
2023-01-10 13:49:33 +00:00
|
|
|
// NOTE: When operating without hardIRQ/softIRQ, applications tends to call ioctl request
|
|
|
|
// for runtime of PCM substream in the interval equivalent to the size of PCM buffer. It
|
|
|
|
// could take a round over queue of AMDTP packet descriptors and small loss of history. For
|
|
|
|
// safe, keep more 8 elements for the queue, equivalent to 1 ms.
|
|
|
|
descs = kcalloc(s->queue_size + 8, sizeof(*descs), GFP_KERNEL);
|
2023-01-09 02:17:36 +00:00
|
|
|
if (!descs) {
|
2019-07-22 03:36:58 +00:00
|
|
|
err = -ENOMEM;
|
|
|
|
goto err_context;
|
|
|
|
}
|
2023-01-09 02:17:38 +00:00
|
|
|
s->packet_descs = descs;
|
2023-01-09 02:17:36 +00:00
|
|
|
|
|
|
|
INIT_LIST_HEAD(&s->packet_descs_list);
|
|
|
|
for (i = 0; i < s->queue_size; ++i) {
|
|
|
|
INIT_LIST_HEAD(&descs->link);
|
|
|
|
list_add_tail(&descs->link, &s->packet_descs_list);
|
|
|
|
++descs;
|
|
|
|
}
|
2023-01-09 02:17:38 +00:00
|
|
|
s->packet_descs_cursor = list_first_entry(&s->packet_descs_list, struct pkt_desc, link);
|
2019-07-22 03:36:58 +00:00
|
|
|
|
2011-03-15 06:57:24 +00:00
|
|
|
s->packet_index = 0;
|
2014-04-25 13:44:45 +00:00
|
|
|
do {
|
2019-05-23 15:14:39 +00:00
|
|
|
struct fw_iso_packet params;
|
2019-10-17 15:54:22 +00:00
|
|
|
|
2019-05-23 15:14:40 +00:00
|
|
|
if (s->direction == AMDTP_IN_STREAM) {
|
2019-10-18 06:19:10 +00:00
|
|
|
err = queue_in_packet(s, ¶ms);
|
2019-05-23 15:14:40 +00:00
|
|
|
} else {
|
2019-10-18 06:19:10 +00:00
|
|
|
bool sched_irq = false;
|
|
|
|
|
2019-05-23 15:14:40 +00:00
|
|
|
params.header_length = 0;
|
|
|
|
params.payload_length = 0;
|
2019-10-18 06:19:10 +00:00
|
|
|
|
|
|
|
if (is_irq_target) {
|
|
|
|
sched_irq = !((s->packet_index + 1) %
|
|
|
|
idle_irq_interval);
|
|
|
|
}
|
|
|
|
|
2019-10-17 15:54:22 +00:00
|
|
|
err = queue_out_packet(s, ¶ms, sched_irq);
|
2019-05-23 15:14:40 +00:00
|
|
|
}
|
2014-04-25 13:44:45 +00:00
|
|
|
if (err < 0)
|
2019-07-22 03:36:58 +00:00
|
|
|
goto err_pkt_descs;
|
2014-04-25 13:44:45 +00:00
|
|
|
} while (s->packet_index > 0);
|
2011-03-15 06:53:21 +00:00
|
|
|
|
2014-04-25 13:44:46 +00:00
|
|
|
/* NOTE: TAG1 matches CIP. This just affects in stream. */
|
2014-04-25 13:45:03 +00:00
|
|
|
tag = FW_ISO_CONTEXT_MATCH_TAG1;
|
ALSA: firewire-lib: add no-header packet processing
As long as investigating Fireface 400, IEC 61883-1/6 is not applied to
its packet streaming protocol. Remarks of the specific protocol are:
* Each packet doesn't include CIP headers.
* 64,0 and 128,0 kHz are supported.
* The device doesn't necessarily transmit 8,000 packets per second.
* 0, 1, 2, 3 are used as tag for rx isochronous packets, however 0 is
used for tx isochronous packets.
On the other hand, there's a common feature. The number of data blocks
transferred in a second is the same as sampling transmission frequency.
Current ALSA IEC 61883-1/6 engine already has a method to calculate it and
this driver can utilize it for rx packets, as well as tx packets.
This commit adds support for the transferring protocol. CIP_NO_HEADERS
flag is newly added. When this flag is set:
* Both of 0 (without CIP header) and 1 (with CIP header) are used as tag
to handle incoming isochronous packet.
* 0 (without CIP header) is used as tag to transfer outgoing isochronous
packet.
* Skip CIP header evaluation.
* Use unique way to calculate the quadlets of isochronous packet payload.
In ALSA PCM interface, 128.0 kHz is not supported, and the ALSA
IEC 61883-1/6 engine doesn't support 64.0 kHz. These modes are dropped.
The sequence of rx packet has a remarkable quirk about tag. This will be
described in later commits.
Signed-off-by: Takashi Sakamoto <o-takashi@sakamocchi.jp>
Signed-off-by: Takashi Iwai <tiwai@suse.de>
2017-03-31 13:06:07 +00:00
|
|
|
if ((s->flags & CIP_EMPTY_WITH_TAG0) || (s->flags & CIP_NO_HEADER))
|
2014-04-25 13:45:03 +00:00
|
|
|
tag |= FW_ISO_CONTEXT_MATCH_TAG0;
|
|
|
|
|
2021-05-20 04:01:54 +00:00
|
|
|
s->ready_processing = false;
|
2021-05-20 04:01:52 +00:00
|
|
|
err = fw_iso_context_start(s->context, -1, 0, tag);
|
2011-03-15 06:53:21 +00:00
|
|
|
if (err < 0)
|
2019-07-22 03:36:58 +00:00
|
|
|
goto err_pkt_descs;
|
2011-03-15 06:53:21 +00:00
|
|
|
|
|
|
|
mutex_unlock(&s->mutex);
|
|
|
|
|
|
|
|
return 0;
|
2019-07-22 03:36:58 +00:00
|
|
|
err_pkt_descs:
|
2023-01-09 02:17:38 +00:00
|
|
|
kfree(s->packet_descs);
|
|
|
|
s->packet_descs = NULL;
|
2011-03-15 06:53:21 +00:00
|
|
|
err_context:
|
2021-05-27 12:26:09 +00:00
|
|
|
if (s->direction == AMDTP_OUT_STREAM) {
|
2021-05-22 01:32:59 +00:00
|
|
|
kfree(s->ctx_data.rx.seq.descs);
|
2021-05-27 12:26:09 +00:00
|
|
|
} else {
|
|
|
|
if (s->domain->replay.enable)
|
|
|
|
kfree(s->ctx_data.tx.cache.descs);
|
|
|
|
}
|
2011-03-15 06:53:21 +00:00
|
|
|
fw_iso_context_destroy(s->context);
|
|
|
|
s->context = ERR_PTR(-1);
|
|
|
|
err_buffer:
|
|
|
|
iso_packets_buffer_destroy(&s->buffer, s->unit);
|
|
|
|
err_unlock:
|
|
|
|
mutex_unlock(&s->mutex);
|
|
|
|
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
2012-05-13 16:49:14 +00:00
|
|
|
/**
|
2019-10-18 06:19:07 +00:00
|
|
|
* amdtp_domain_stream_pcm_pointer - get the PCM buffer position
|
|
|
|
* @d: the AMDTP domain.
|
2014-04-25 13:44:42 +00:00
|
|
|
* @s: the AMDTP stream that transports the PCM data
|
2012-05-13 16:49:14 +00:00
|
|
|
*
|
|
|
|
* Returns the current buffer position, in frames.
|
|
|
|
*/
|
2019-10-18 06:19:07 +00:00
|
|
|
unsigned long amdtp_domain_stream_pcm_pointer(struct amdtp_domain *d,
|
|
|
|
struct amdtp_stream *s)
|
2012-05-13 16:49:14 +00:00
|
|
|
{
|
2019-10-18 06:19:07 +00:00
|
|
|
struct amdtp_stream *irq_target = d->irq_target;
|
|
|
|
|
|
|
|
if (irq_target && amdtp_stream_running(irq_target)) {
|
2024-09-04 12:51:54 +00:00
|
|
|
// The work item to call snd_pcm_period_elapsed() can reach here by the call of
|
|
|
|
// snd_pcm_ops.pointer(), however less packets would be available then. Therefore
|
|
|
|
// the following call is just for user process contexts.
|
2024-07-30 19:53:29 +00:00
|
|
|
if (current_work() != &s->period_work)
|
2019-10-18 06:19:07 +00:00
|
|
|
fw_iso_context_flush_completions(irq_target->context);
|
|
|
|
}
|
2012-05-13 16:49:14 +00:00
|
|
|
|
locking/atomics: COCCINELLE/treewide: Convert trivial ACCESS_ONCE() patterns to READ_ONCE()/WRITE_ONCE()
Please do not apply this to mainline directly, instead please re-run the
coccinelle script shown below and apply its output.
For several reasons, it is desirable to use {READ,WRITE}_ONCE() in
preference to ACCESS_ONCE(), and new code is expected to use one of the
former. So far, there's been no reason to change most existing uses of
ACCESS_ONCE(), as these aren't harmful, and changing them results in
churn.
However, for some features, the read/write distinction is critical to
correct operation. To distinguish these cases, separate read/write
accessors must be used. This patch migrates (most) remaining
ACCESS_ONCE() instances to {READ,WRITE}_ONCE(), using the following
coccinelle script:
----
// Convert trivial ACCESS_ONCE() uses to equivalent READ_ONCE() and
// WRITE_ONCE()
// $ make coccicheck COCCI=/home/mark/once.cocci SPFLAGS="--include-headers" MODE=patch
virtual patch
@ depends on patch @
expression E1, E2;
@@
- ACCESS_ONCE(E1) = E2
+ WRITE_ONCE(E1, E2)
@ depends on patch @
expression E;
@@
- ACCESS_ONCE(E)
+ READ_ONCE(E)
----
Signed-off-by: Mark Rutland <mark.rutland@arm.com>
Signed-off-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: davem@davemloft.net
Cc: linux-arch@vger.kernel.org
Cc: mpe@ellerman.id.au
Cc: shuah@kernel.org
Cc: snitzer@redhat.com
Cc: thor.thayer@linux.intel.com
Cc: tj@kernel.org
Cc: viro@zeniv.linux.org.uk
Cc: will.deacon@arm.com
Link: http://lkml.kernel.org/r/1508792849-3115-19-git-send-email-paulmck@linux.vnet.ibm.com
Signed-off-by: Ingo Molnar <mingo@kernel.org>
2017-10-23 21:07:29 +00:00
|
|
|
return READ_ONCE(s->pcm_buffer_pointer);
|
2012-05-13 16:49:14 +00:00
|
|
|
}
|
2019-10-18 06:19:07 +00:00
|
|
|
EXPORT_SYMBOL_GPL(amdtp_domain_stream_pcm_pointer);
|
2012-05-13 16:49:14 +00:00
|
|
|
|
2017-06-07 00:38:05 +00:00
|
|
|
/**
|
2019-10-18 06:19:08 +00:00
|
|
|
* amdtp_domain_stream_pcm_ack - acknowledge queued PCM frames
|
|
|
|
* @d: the AMDTP domain.
|
2017-06-07 00:38:05 +00:00
|
|
|
* @s: the AMDTP stream that transfers the PCM frames
|
|
|
|
*
|
|
|
|
* Returns zero always.
|
|
|
|
*/
|
2019-10-18 06:19:08 +00:00
|
|
|
int amdtp_domain_stream_pcm_ack(struct amdtp_domain *d, struct amdtp_stream *s)
|
2017-06-07 00:38:05 +00:00
|
|
|
{
|
2019-10-18 06:19:08 +00:00
|
|
|
struct amdtp_stream *irq_target = d->irq_target;
|
|
|
|
|
|
|
|
// Process isochronous packets for recent isochronous cycle to handle
|
|
|
|
// queued PCM frames.
|
2021-06-06 02:56:51 +00:00
|
|
|
if (irq_target && amdtp_stream_running(irq_target))
|
2019-10-18 06:19:08 +00:00
|
|
|
fw_iso_context_flush_completions(irq_target->context);
|
2017-06-07 00:38:05 +00:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
2019-10-18 06:19:08 +00:00
|
|
|
EXPORT_SYMBOL_GPL(amdtp_domain_stream_pcm_ack);
|
2017-06-07 00:38:05 +00:00
|
|
|
|
2011-03-15 06:53:21 +00:00
|
|
|
/**
|
2014-04-25 13:44:42 +00:00
|
|
|
* amdtp_stream_update - update the stream after a bus reset
|
|
|
|
* @s: the AMDTP stream
|
2011-03-15 06:53:21 +00:00
|
|
|
*/
|
2014-04-25 13:44:42 +00:00
|
|
|
void amdtp_stream_update(struct amdtp_stream *s)
|
2011-03-15 06:53:21 +00:00
|
|
|
{
|
2015-05-22 14:21:12 +00:00
|
|
|
/* Precomputing. */
|
locking/atomics: COCCINELLE/treewide: Convert trivial ACCESS_ONCE() patterns to READ_ONCE()/WRITE_ONCE()
Please do not apply this to mainline directly, instead please re-run the
coccinelle script shown below and apply its output.
For several reasons, it is desirable to use {READ,WRITE}_ONCE() in
preference to ACCESS_ONCE(), and new code is expected to use one of the
former. So far, there's been no reason to change most existing uses of
ACCESS_ONCE(), as these aren't harmful, and changing them results in
churn.
However, for some features, the read/write distinction is critical to
correct operation. To distinguish these cases, separate read/write
accessors must be used. This patch migrates (most) remaining
ACCESS_ONCE() instances to {READ,WRITE}_ONCE(), using the following
coccinelle script:
----
// Convert trivial ACCESS_ONCE() uses to equivalent READ_ONCE() and
// WRITE_ONCE()
// $ make coccicheck COCCI=/home/mark/once.cocci SPFLAGS="--include-headers" MODE=patch
virtual patch
@ depends on patch @
expression E1, E2;
@@
- ACCESS_ONCE(E1) = E2
+ WRITE_ONCE(E1, E2)
@ depends on patch @
expression E;
@@
- ACCESS_ONCE(E)
+ READ_ONCE(E)
----
Signed-off-by: Mark Rutland <mark.rutland@arm.com>
Signed-off-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: davem@davemloft.net
Cc: linux-arch@vger.kernel.org
Cc: mpe@ellerman.id.au
Cc: shuah@kernel.org
Cc: snitzer@redhat.com
Cc: thor.thayer@linux.intel.com
Cc: tj@kernel.org
Cc: viro@zeniv.linux.org.uk
Cc: will.deacon@arm.com
Link: http://lkml.kernel.org/r/1508792849-3115-19-git-send-email-paulmck@linux.vnet.ibm.com
Signed-off-by: Ingo Molnar <mingo@kernel.org>
2017-10-23 21:07:29 +00:00
|
|
|
WRITE_ONCE(s->source_node_id_field,
|
|
|
|
(fw_parent_device(s->unit)->card->node_id << CIP_SID_SHIFT) & CIP_SID_MASK);
|
2011-03-15 06:53:21 +00:00
|
|
|
}
|
2014-04-25 13:44:42 +00:00
|
|
|
EXPORT_SYMBOL(amdtp_stream_update);
|
2011-03-15 06:53:21 +00:00
|
|
|
|
|
|
|
/**
|
2014-04-25 13:44:42 +00:00
|
|
|
* amdtp_stream_stop - stop sending packets
|
|
|
|
* @s: the AMDTP stream to stop
|
2011-03-15 06:53:21 +00:00
|
|
|
*
|
|
|
|
* All PCM and MIDI devices of the stream must be stopped before the stream
|
|
|
|
* itself can be stopped.
|
|
|
|
*/
|
2019-08-04 06:21:38 +00:00
|
|
|
static void amdtp_stream_stop(struct amdtp_stream *s)
|
2011-03-15 06:53:21 +00:00
|
|
|
{
|
|
|
|
mutex_lock(&s->mutex);
|
|
|
|
|
2014-04-25 13:44:42 +00:00
|
|
|
if (!amdtp_stream_running(s)) {
|
2011-03-15 06:53:21 +00:00
|
|
|
mutex_unlock(&s->mutex);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2024-07-30 19:53:26 +00:00
|
|
|
cancel_work_sync(&s->period_work);
|
2011-03-15 06:53:21 +00:00
|
|
|
fw_iso_context_stop(s->context);
|
|
|
|
fw_iso_context_destroy(s->context);
|
|
|
|
s->context = ERR_PTR(-1);
|
|
|
|
iso_packets_buffer_destroy(&s->buffer, s->unit);
|
2023-01-09 02:17:38 +00:00
|
|
|
kfree(s->packet_descs);
|
|
|
|
s->packet_descs = NULL;
|
2011-03-15 06:53:21 +00:00
|
|
|
|
2021-05-27 12:26:09 +00:00
|
|
|
if (s->direction == AMDTP_OUT_STREAM) {
|
2021-05-22 01:32:59 +00:00
|
|
|
kfree(s->ctx_data.rx.seq.descs);
|
2021-05-27 12:26:09 +00:00
|
|
|
} else {
|
|
|
|
if (s->domain->replay.enable)
|
|
|
|
kfree(s->ctx_data.tx.cache.descs);
|
|
|
|
}
|
2014-04-25 13:44:49 +00:00
|
|
|
|
2011-03-15 06:53:21 +00:00
|
|
|
mutex_unlock(&s->mutex);
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
2014-04-25 13:44:42 +00:00
|
|
|
* amdtp_stream_pcm_abort - abort the running PCM device
|
2011-03-15 06:53:21 +00:00
|
|
|
* @s: the AMDTP stream about to be stopped
|
|
|
|
*
|
|
|
|
* If the isochronous stream needs to be stopped asynchronously, call this
|
|
|
|
* function first to stop the PCM device.
|
|
|
|
*/
|
2014-04-25 13:44:42 +00:00
|
|
|
void amdtp_stream_pcm_abort(struct amdtp_stream *s)
|
2011-03-15 06:53:21 +00:00
|
|
|
{
|
|
|
|
struct snd_pcm_substream *pcm;
|
|
|
|
|
locking/atomics: COCCINELLE/treewide: Convert trivial ACCESS_ONCE() patterns to READ_ONCE()/WRITE_ONCE()
Please do not apply this to mainline directly, instead please re-run the
coccinelle script shown below and apply its output.
For several reasons, it is desirable to use {READ,WRITE}_ONCE() in
preference to ACCESS_ONCE(), and new code is expected to use one of the
former. So far, there's been no reason to change most existing uses of
ACCESS_ONCE(), as these aren't harmful, and changing them results in
churn.
However, for some features, the read/write distinction is critical to
correct operation. To distinguish these cases, separate read/write
accessors must be used. This patch migrates (most) remaining
ACCESS_ONCE() instances to {READ,WRITE}_ONCE(), using the following
coccinelle script:
----
// Convert trivial ACCESS_ONCE() uses to equivalent READ_ONCE() and
// WRITE_ONCE()
// $ make coccicheck COCCI=/home/mark/once.cocci SPFLAGS="--include-headers" MODE=patch
virtual patch
@ depends on patch @
expression E1, E2;
@@
- ACCESS_ONCE(E1) = E2
+ WRITE_ONCE(E1, E2)
@ depends on patch @
expression E;
@@
- ACCESS_ONCE(E)
+ READ_ONCE(E)
----
Signed-off-by: Mark Rutland <mark.rutland@arm.com>
Signed-off-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: davem@davemloft.net
Cc: linux-arch@vger.kernel.org
Cc: mpe@ellerman.id.au
Cc: shuah@kernel.org
Cc: snitzer@redhat.com
Cc: thor.thayer@linux.intel.com
Cc: tj@kernel.org
Cc: viro@zeniv.linux.org.uk
Cc: will.deacon@arm.com
Link: http://lkml.kernel.org/r/1508792849-3115-19-git-send-email-paulmck@linux.vnet.ibm.com
Signed-off-by: Ingo Molnar <mingo@kernel.org>
2017-10-23 21:07:29 +00:00
|
|
|
pcm = READ_ONCE(s->pcm);
|
2014-11-07 16:08:28 +00:00
|
|
|
if (pcm)
|
|
|
|
snd_pcm_stop_xrun(pcm);
|
2011-03-15 06:53:21 +00:00
|
|
|
}
|
2014-04-25 13:44:42 +00:00
|
|
|
EXPORT_SYMBOL(amdtp_stream_pcm_abort);
|
2019-08-04 06:21:20 +00:00
|
|
|
|
|
|
|
/**
|
|
|
|
* amdtp_domain_init - initialize an AMDTP domain structure
|
|
|
|
* @d: the AMDTP domain to initialize.
|
|
|
|
*/
|
|
|
|
int amdtp_domain_init(struct amdtp_domain *d)
|
|
|
|
{
|
|
|
|
INIT_LIST_HEAD(&d->streams);
|
|
|
|
|
2019-10-07 11:05:16 +00:00
|
|
|
d->events_per_period = 0;
|
|
|
|
|
2019-08-04 06:21:20 +00:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(amdtp_domain_init);
|
|
|
|
|
|
|
|
/**
|
|
|
|
* amdtp_domain_destroy - destroy an AMDTP domain structure
|
|
|
|
* @d: the AMDTP domain to destroy.
|
|
|
|
*/
|
|
|
|
void amdtp_domain_destroy(struct amdtp_domain *d)
|
|
|
|
{
|
2019-09-06 13:14:14 +00:00
|
|
|
// At present nothing to do.
|
|
|
|
return;
|
2019-08-04 06:21:20 +00:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(amdtp_domain_destroy);
|
2019-08-04 06:21:21 +00:00
|
|
|
|
2019-08-04 06:21:22 +00:00
|
|
|
/**
|
|
|
|
* amdtp_domain_add_stream - register isoc context into the domain.
|
|
|
|
* @d: the AMDTP domain.
|
|
|
|
* @s: the AMDTP stream.
|
|
|
|
* @channel: the isochronous channel on the bus.
|
|
|
|
* @speed: firewire speed code.
|
|
|
|
*/
|
|
|
|
int amdtp_domain_add_stream(struct amdtp_domain *d, struct amdtp_stream *s,
|
|
|
|
int channel, int speed)
|
|
|
|
{
|
|
|
|
struct amdtp_stream *tmp;
|
|
|
|
|
|
|
|
list_for_each_entry(tmp, &d->streams, list) {
|
|
|
|
if (s == tmp)
|
|
|
|
return -EBUSY;
|
|
|
|
}
|
|
|
|
|
|
|
|
list_add(&s->list, &d->streams);
|
|
|
|
|
|
|
|
s->channel = channel;
|
|
|
|
s->speed = speed;
|
2020-05-08 04:36:28 +00:00
|
|
|
s->domain = d;
|
2019-08-04 06:21:22 +00:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(amdtp_domain_add_stream);
|
|
|
|
|
ALSA: firewire-lib: replay sequence of incoming packets for outgoing packets
ALSA IEC 61883-1/6 packet streaming engine uses pre-computed parameters
ideal for nominal sampling transfer frequency (STF) to transfer packets
to device since it was added 2011. As a result of user experience for a
decade, it is clear that the sequence is not suitable to some actual
devices. It takes the devices to generate noise, and causes any type of
discontinuity in the series of packet transferred from the device. It's
required for the engine to transfer packets according to effective STF.
The effective STF is given by media clock recovered by the sequence of
packet transferred from the target device. In the previous commit, the
sequence is already cached. The media clock recovery can be achieved by
analyzing the sequence.
In technological world, many ideas are proposed for media clock recovery.
However, the small part of them could be actually adopted in our case
since floating point arithmetic is not mostly available in Linux kernel
land.
This commit adopts the simple way from them; sequence replay, which means
that the sequence of parameters from incoming packet is used as is to
transfer outgoing packets. The media clock is not computed internally,
but the sequence of outgoing packet superficially looks to be generated by
the media clock.
The association between source and destination is decided when starting
AMDTP domain. When the target device supports a pair of isochronous packet
streams, the tx stream is source and the rx stream is destination. When it
supports two pair of streams, each of tx stream is associated to
corresponding rx stream in its order. When it supports less number of tx
streams than rx streams, the fist tx stream is selected for all of rx
streams. When it supports more tx streams than rx streams, the first tx
packet is associated to the rx stream.
As I noted in previous commit, the sequence of parameters from incoming
packet is different between devices, time to time. It is worse idea to
replay the sequence of parameters from a device for the sequence of
packet to the other devices even if they are in the same category of
device.
Signed-off-by: Takashi Sakamoto <o-takashi@sakamocchi.jp>
Link: https://lore.kernel.org/r/20210527122611.173711-3-o-takashi@sakamocchi.jp
Signed-off-by: Takashi Iwai <tiwai@suse.de>
2021-05-27 12:26:10 +00:00
|
|
|
// Make the reference from rx stream to tx stream for sequence replay. When the number of tx streams
|
|
|
|
// is less than the number of rx streams, the first tx stream is selected.
|
|
|
|
static int make_association(struct amdtp_domain *d)
|
|
|
|
{
|
|
|
|
unsigned int dst_index = 0;
|
|
|
|
struct amdtp_stream *rx;
|
|
|
|
|
|
|
|
// Make association to replay target.
|
|
|
|
list_for_each_entry(rx, &d->streams, list) {
|
|
|
|
if (rx->direction == AMDTP_OUT_STREAM) {
|
|
|
|
unsigned int src_index = 0;
|
|
|
|
struct amdtp_stream *tx = NULL;
|
|
|
|
struct amdtp_stream *s;
|
|
|
|
|
|
|
|
list_for_each_entry(s, &d->streams, list) {
|
|
|
|
if (s->direction == AMDTP_IN_STREAM) {
|
|
|
|
if (dst_index == src_index) {
|
|
|
|
tx = s;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
++src_index;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (!tx) {
|
|
|
|
// Select the first entry.
|
|
|
|
list_for_each_entry(s, &d->streams, list) {
|
|
|
|
if (s->direction == AMDTP_IN_STREAM) {
|
|
|
|
tx = s;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// No target is available to replay sequence.
|
|
|
|
if (!tx)
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
rx->ctx_data.rx.replay_target = tx;
|
|
|
|
|
|
|
|
++dst_index;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2019-08-04 06:21:23 +00:00
|
|
|
/**
|
|
|
|
* amdtp_domain_start - start sending packets for isoc context in the domain.
|
|
|
|
* @d: the AMDTP domain.
|
2021-05-20 04:01:51 +00:00
|
|
|
* @tx_init_skip_cycles: the number of cycles to skip processing packets at initial stage of IR
|
|
|
|
* contexts.
|
2021-05-27 12:26:09 +00:00
|
|
|
* @replay_seq: whether to replay the sequence of packet in IR context for the sequence of packet in
|
|
|
|
* IT context.
|
2021-05-27 12:26:11 +00:00
|
|
|
* @replay_on_the_fly: transfer rx packets according to nominal frequency, then begin to replay
|
|
|
|
* according to arrival of events in tx packets.
|
2019-08-04 06:21:23 +00:00
|
|
|
*/
|
2021-05-27 12:26:11 +00:00
|
|
|
int amdtp_domain_start(struct amdtp_domain *d, unsigned int tx_init_skip_cycles, bool replay_seq,
|
|
|
|
bool replay_on_the_fly)
|
2019-08-04 06:21:23 +00:00
|
|
|
{
|
2020-05-08 04:36:29 +00:00
|
|
|
unsigned int events_per_buffer = d->events_per_buffer;
|
|
|
|
unsigned int events_per_period = d->events_per_period;
|
|
|
|
unsigned int queue_size;
|
2019-08-04 06:21:23 +00:00
|
|
|
struct amdtp_stream *s;
|
2021-06-24 18:49:36 +00:00
|
|
|
bool found = false;
|
2019-10-18 06:19:11 +00:00
|
|
|
int err;
|
2019-08-04 06:21:23 +00:00
|
|
|
|
ALSA: firewire-lib: replay sequence of incoming packets for outgoing packets
ALSA IEC 61883-1/6 packet streaming engine uses pre-computed parameters
ideal for nominal sampling transfer frequency (STF) to transfer packets
to device since it was added 2011. As a result of user experience for a
decade, it is clear that the sequence is not suitable to some actual
devices. It takes the devices to generate noise, and causes any type of
discontinuity in the series of packet transferred from the device. It's
required for the engine to transfer packets according to effective STF.
The effective STF is given by media clock recovered by the sequence of
packet transferred from the target device. In the previous commit, the
sequence is already cached. The media clock recovery can be achieved by
analyzing the sequence.
In technological world, many ideas are proposed for media clock recovery.
However, the small part of them could be actually adopted in our case
since floating point arithmetic is not mostly available in Linux kernel
land.
This commit adopts the simple way from them; sequence replay, which means
that the sequence of parameters from incoming packet is used as is to
transfer outgoing packets. The media clock is not computed internally,
but the sequence of outgoing packet superficially looks to be generated by
the media clock.
The association between source and destination is decided when starting
AMDTP domain. When the target device supports a pair of isochronous packet
streams, the tx stream is source and the rx stream is destination. When it
supports two pair of streams, each of tx stream is associated to
corresponding rx stream in its order. When it supports less number of tx
streams than rx streams, the fist tx stream is selected for all of rx
streams. When it supports more tx streams than rx streams, the first tx
packet is associated to the rx stream.
As I noted in previous commit, the sequence of parameters from incoming
packet is different between devices, time to time. It is worse idea to
replay the sequence of parameters from a device for the sequence of
packet to the other devices even if they are in the same category of
device.
Signed-off-by: Takashi Sakamoto <o-takashi@sakamocchi.jp>
Link: https://lore.kernel.org/r/20210527122611.173711-3-o-takashi@sakamocchi.jp
Signed-off-by: Takashi Iwai <tiwai@suse.de>
2021-05-27 12:26:10 +00:00
|
|
|
if (replay_seq) {
|
|
|
|
err = make_association(d);
|
|
|
|
if (err < 0)
|
|
|
|
return err;
|
|
|
|
}
|
2021-05-27 12:26:09 +00:00
|
|
|
d->replay.enable = replay_seq;
|
2021-05-27 12:26:11 +00:00
|
|
|
d->replay.on_the_fly = replay_on_the_fly;
|
2021-05-27 12:26:09 +00:00
|
|
|
|
2019-10-18 06:19:10 +00:00
|
|
|
// Select an IT context as IRQ target.
|
2019-08-04 06:21:23 +00:00
|
|
|
list_for_each_entry(s, &d->streams, list) {
|
2021-06-24 18:49:36 +00:00
|
|
|
if (s->direction == AMDTP_OUT_STREAM) {
|
|
|
|
found = true;
|
2019-08-04 06:21:23 +00:00
|
|
|
break;
|
2021-06-24 18:49:36 +00:00
|
|
|
}
|
2019-08-04 06:21:23 +00:00
|
|
|
}
|
2021-06-24 18:49:36 +00:00
|
|
|
if (!found)
|
2019-10-18 06:19:10 +00:00
|
|
|
return -ENXIO;
|
|
|
|
d->irq_target = s;
|
2019-08-04 06:21:23 +00:00
|
|
|
|
2021-05-20 04:01:51 +00:00
|
|
|
d->processing_cycle.tx_init_skip = tx_init_skip_cycles;
|
|
|
|
|
2020-05-08 04:36:29 +00:00
|
|
|
// This is a case that AMDTP streams in domain run just for MIDI
|
|
|
|
// substream. Use the number of events equivalent to 10 msec as
|
|
|
|
// interval of hardware IRQ.
|
|
|
|
if (events_per_period == 0)
|
|
|
|
events_per_period = amdtp_rate_table[d->irq_target->sfc] / 100;
|
|
|
|
if (events_per_buffer == 0)
|
|
|
|
events_per_buffer = events_per_period * 3;
|
|
|
|
|
|
|
|
queue_size = DIV_ROUND_UP(CYCLES_PER_SECOND * events_per_buffer,
|
|
|
|
amdtp_rate_table[d->irq_target->sfc]);
|
|
|
|
|
2019-10-18 06:19:10 +00:00
|
|
|
list_for_each_entry(s, &d->streams, list) {
|
2021-05-20 04:01:52 +00:00
|
|
|
unsigned int idle_irq_interval = 0;
|
2019-10-18 06:19:11 +00:00
|
|
|
|
2021-05-20 04:01:52 +00:00
|
|
|
if (s->direction == AMDTP_OUT_STREAM && s == d->irq_target) {
|
|
|
|
idle_irq_interval = DIV_ROUND_UP(CYCLES_PER_SECOND * events_per_period,
|
|
|
|
amdtp_rate_table[d->irq_target->sfc]);
|
2019-10-18 06:19:10 +00:00
|
|
|
}
|
2019-08-04 06:21:23 +00:00
|
|
|
|
2021-05-20 04:01:52 +00:00
|
|
|
// Starts immediately but actually DMA context starts several hundred cycles later.
|
|
|
|
err = amdtp_stream_start(s, s->channel, s->speed, queue_size, idle_irq_interval);
|
|
|
|
if (err < 0)
|
|
|
|
goto error;
|
|
|
|
}
|
2019-10-18 06:19:10 +00:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
error:
|
|
|
|
list_for_each_entry(s, &d->streams, list)
|
|
|
|
amdtp_stream_stop(s);
|
2019-08-04 06:21:23 +00:00
|
|
|
return err;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(amdtp_domain_start);
|
|
|
|
|
2019-08-04 06:21:21 +00:00
|
|
|
/**
|
|
|
|
* amdtp_domain_stop - stop sending packets for isoc context in the same domain.
|
|
|
|
* @d: the AMDTP domain to which the isoc contexts belong.
|
|
|
|
*/
|
|
|
|
void amdtp_domain_stop(struct amdtp_domain *d)
|
|
|
|
{
|
|
|
|
struct amdtp_stream *s, *next;
|
|
|
|
|
2019-10-18 06:19:10 +00:00
|
|
|
if (d->irq_target)
|
|
|
|
amdtp_stream_stop(d->irq_target);
|
|
|
|
|
2019-08-04 06:21:21 +00:00
|
|
|
list_for_each_entry_safe(s, next, &d->streams, list) {
|
|
|
|
list_del(&s->list);
|
|
|
|
|
2019-10-18 06:19:10 +00:00
|
|
|
if (s != d->irq_target)
|
|
|
|
amdtp_stream_stop(s);
|
2019-08-04 06:21:21 +00:00
|
|
|
}
|
2019-10-07 11:05:16 +00:00
|
|
|
|
|
|
|
d->events_per_period = 0;
|
2019-10-18 06:19:10 +00:00
|
|
|
d->irq_target = NULL;
|
2019-08-04 06:21:21 +00:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(amdtp_domain_stop);
|