forked from luck/tmp_suning_uos_patched
firewire: Use a buffer fill descriptor for receive when header size is 0.
When the DMA is setup to not strip any headers, we need to use the buffer fill descriptor instead of the dual buffer, since the dual buffer descriptor must strip a non-zero number of header quadlets. Signed-off-by: Kristian Høgsberg <krh@redhat.com> Signed-off-by: Stefan Richter <stefanr@s5r6.in-berlin.de>
This commit is contained in:
parent
1e1d196bf5
commit
d2746dc192
@ -111,7 +111,7 @@ fw_iso_context_create(struct fw_card *card, int type, size_t header_size,
|
|||||||
{
|
{
|
||||||
struct fw_iso_context *ctx;
|
struct fw_iso_context *ctx;
|
||||||
|
|
||||||
ctx = card->driver->allocate_iso_context(card, type);
|
ctx = card->driver->allocate_iso_context(card, type, header_size);
|
||||||
if (IS_ERR(ctx))
|
if (IS_ERR(ctx))
|
||||||
return ctx;
|
return ctx;
|
||||||
|
|
||||||
|
@ -1267,16 +1267,34 @@ ohci_enable_phys_dma(struct fw_card *card, int node_id, int generation)
|
|||||||
spin_unlock_irqrestore(&ohci->lock, flags);
|
spin_unlock_irqrestore(&ohci->lock, flags);
|
||||||
return retval;
|
return retval;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static int handle_ir_bufferfill_packet(struct context *context,
|
||||||
|
struct descriptor *d,
|
||||||
|
struct descriptor *last)
|
||||||
|
{
|
||||||
|
struct iso_context *ctx =
|
||||||
|
container_of(context, struct iso_context, context);
|
||||||
|
|
||||||
static int handle_ir_packet(struct context *context,
|
if (d->res_count > 0)
|
||||||
struct descriptor *d,
|
return 0;
|
||||||
struct descriptor *last)
|
|
||||||
|
if (le16_to_cpu(last->control) & descriptor_irq_always)
|
||||||
|
ctx->base.callback(&ctx->base,
|
||||||
|
le16_to_cpu(last->res_count),
|
||||||
|
0, NULL, ctx->base.callback_data);
|
||||||
|
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
static int handle_ir_dualbuffer_packet(struct context *context,
|
||||||
|
struct descriptor *d,
|
||||||
|
struct descriptor *last)
|
||||||
{
|
{
|
||||||
struct iso_context *ctx =
|
struct iso_context *ctx =
|
||||||
container_of(context, struct iso_context, context);
|
container_of(context, struct iso_context, context);
|
||||||
struct db_descriptor *db = (struct db_descriptor *) d;
|
struct db_descriptor *db = (struct db_descriptor *) d;
|
||||||
size_t header_length;
|
size_t header_length;
|
||||||
|
|
||||||
if (db->first_res_count > 0 && db->second_res_count > 0)
|
if (db->first_res_count > 0 && db->second_res_count > 0)
|
||||||
/* This descriptor isn't done yet, stop iteration. */
|
/* This descriptor isn't done yet, stop iteration. */
|
||||||
return 0;
|
return 0;
|
||||||
@ -1317,7 +1335,7 @@ static int handle_it_packet(struct context *context,
|
|||||||
}
|
}
|
||||||
|
|
||||||
static struct fw_iso_context *
|
static struct fw_iso_context *
|
||||||
ohci_allocate_iso_context(struct fw_card *card, int type)
|
ohci_allocate_iso_context(struct fw_card *card, int type, size_t header_size)
|
||||||
{
|
{
|
||||||
struct fw_ohci *ohci = fw_ohci(card);
|
struct fw_ohci *ohci = fw_ohci(card);
|
||||||
struct iso_context *ctx, *list;
|
struct iso_context *ctx, *list;
|
||||||
@ -1333,7 +1351,10 @@ ohci_allocate_iso_context(struct fw_card *card, int type)
|
|||||||
} else {
|
} else {
|
||||||
mask = &ohci->ir_context_mask;
|
mask = &ohci->ir_context_mask;
|
||||||
list = ohci->ir_context_list;
|
list = ohci->ir_context_list;
|
||||||
callback = handle_ir_packet;
|
if (header_size > 0)
|
||||||
|
callback = handle_ir_dualbuffer_packet;
|
||||||
|
else
|
||||||
|
callback = handle_ir_bufferfill_packet;
|
||||||
}
|
}
|
||||||
|
|
||||||
spin_lock_irqsave(&ohci->lock, flags);
|
spin_lock_irqsave(&ohci->lock, flags);
|
||||||
@ -1378,7 +1399,7 @@ static int ohci_start_iso(struct fw_iso_context *base, s32 cycle)
|
|||||||
{
|
{
|
||||||
struct iso_context *ctx = container_of(base, struct iso_context, base);
|
struct iso_context *ctx = container_of(base, struct iso_context, base);
|
||||||
struct fw_ohci *ohci = ctx->context.ohci;
|
struct fw_ohci *ohci = ctx->context.ohci;
|
||||||
u32 cycle_match = 0;
|
u32 cycle_match = 0, mode;
|
||||||
int index;
|
int index;
|
||||||
|
|
||||||
if (ctx->base.type == FW_ISO_CONTEXT_TRANSMIT) {
|
if (ctx->base.type == FW_ISO_CONTEXT_TRANSMIT) {
|
||||||
@ -1393,11 +1414,15 @@ static int ohci_start_iso(struct fw_iso_context *base, s32 cycle)
|
|||||||
} else {
|
} else {
|
||||||
index = ctx - ohci->ir_context_list;
|
index = ctx - ohci->ir_context_list;
|
||||||
|
|
||||||
|
if (ctx->base.header_size > 0)
|
||||||
|
mode = IR_CONTEXT_DUAL_BUFFER_MODE;
|
||||||
|
else
|
||||||
|
mode = IR_CONTEXT_BUFFER_FILL;
|
||||||
reg_write(ohci, OHCI1394_IsoRecvIntEventClear, 1 << index);
|
reg_write(ohci, OHCI1394_IsoRecvIntEventClear, 1 << index);
|
||||||
reg_write(ohci, OHCI1394_IsoRecvIntMaskSet, 1 << index);
|
reg_write(ohci, OHCI1394_IsoRecvIntMaskSet, 1 << index);
|
||||||
reg_write(ohci, context_match(ctx->context.regs),
|
reg_write(ohci, context_match(ctx->context.regs),
|
||||||
0xf0000000 | ctx->base.channel);
|
0xf0000000 | ctx->base.channel);
|
||||||
context_run(&ctx->context, IR_CONTEXT_DUAL_BUFFER_MODE);
|
context_run(&ctx->context, mode);
|
||||||
}
|
}
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
@ -1544,10 +1569,10 @@ ohci_queue_iso_transmit(struct fw_iso_context *base,
|
|||||||
}
|
}
|
||||||
|
|
||||||
static int
|
static int
|
||||||
ohci_queue_iso_receive(struct fw_iso_context *base,
|
ohci_queue_iso_receive_dualbuffer(struct fw_iso_context *base,
|
||||||
struct fw_iso_packet *packet,
|
struct fw_iso_packet *packet,
|
||||||
struct fw_iso_buffer *buffer,
|
struct fw_iso_buffer *buffer,
|
||||||
unsigned long payload)
|
unsigned long payload)
|
||||||
{
|
{
|
||||||
struct iso_context *ctx = container_of(base, struct iso_context, base);
|
struct iso_context *ctx = container_of(base, struct iso_context, base);
|
||||||
struct db_descriptor *db = NULL;
|
struct db_descriptor *db = NULL;
|
||||||
@ -1604,10 +1629,58 @@ ohci_queue_iso_receive(struct fw_iso_context *base,
|
|||||||
|
|
||||||
if (p->interrupt)
|
if (p->interrupt)
|
||||||
db->control |= cpu_to_le16(descriptor_irq_always);
|
db->control |= cpu_to_le16(descriptor_irq_always);
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
return 0;
|
static int
|
||||||
}
|
ohci_queue_iso_receive_bufferfill(struct fw_iso_context *base,
|
||||||
|
struct fw_iso_packet *packet,
|
||||||
|
struct fw_iso_buffer *buffer,
|
||||||
|
unsigned long payload)
|
||||||
|
{
|
||||||
|
struct iso_context *ctx = container_of(base, struct iso_context, base);
|
||||||
|
struct descriptor *d = NULL;
|
||||||
|
dma_addr_t d_bus, page_bus;
|
||||||
|
u32 length, rest;
|
||||||
|
int page, offset;
|
||||||
|
|
||||||
|
page = payload >> PAGE_SHIFT;
|
||||||
|
offset = payload & ~PAGE_MASK;
|
||||||
|
rest = packet->payload_length;
|
||||||
|
|
||||||
|
while (rest > 0) {
|
||||||
|
d = context_get_descriptors(&ctx->context, 1, &d_bus);
|
||||||
|
if (d == NULL)
|
||||||
|
return -ENOMEM;
|
||||||
|
|
||||||
|
d->control = cpu_to_le16(descriptor_input_more |
|
||||||
|
descriptor_status |
|
||||||
|
descriptor_branch_always);
|
||||||
|
|
||||||
|
if (offset + rest < PAGE_SIZE)
|
||||||
|
length = rest;
|
||||||
|
else
|
||||||
|
length = PAGE_SIZE - offset;
|
||||||
|
|
||||||
|
page_bus = page_private(buffer->pages[page]);
|
||||||
|
d->data_address = cpu_to_le32(page_bus + offset);
|
||||||
|
d->req_count = cpu_to_le16(length);
|
||||||
|
d->res_count = cpu_to_le16(length);
|
||||||
|
|
||||||
|
context_append(&ctx->context, d, 1, 0);
|
||||||
|
|
||||||
|
offset = (offset + length) & ~PAGE_MASK;
|
||||||
|
rest -= length;
|
||||||
|
page++;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (packet->interrupt)
|
||||||
|
d->control |= cpu_to_le16(descriptor_irq_always);
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
static int
|
static int
|
||||||
ohci_queue_iso(struct fw_iso_context *base,
|
ohci_queue_iso(struct fw_iso_context *base,
|
||||||
struct fw_iso_packet *packet,
|
struct fw_iso_packet *packet,
|
||||||
@ -1616,8 +1689,12 @@ ohci_queue_iso(struct fw_iso_context *base,
|
|||||||
{
|
{
|
||||||
if (base->type == FW_ISO_CONTEXT_TRANSMIT)
|
if (base->type == FW_ISO_CONTEXT_TRANSMIT)
|
||||||
return ohci_queue_iso_transmit(base, packet, buffer, payload);
|
return ohci_queue_iso_transmit(base, packet, buffer, payload);
|
||||||
|
else if (base->header_size == 0)
|
||||||
|
return ohci_queue_iso_receive_bufferfill(base, packet,
|
||||||
|
buffer, payload);
|
||||||
else
|
else
|
||||||
return ohci_queue_iso_receive(base, packet, buffer, payload);
|
return ohci_queue_iso_receive_dualbuffer(base, packet,
|
||||||
|
buffer, payload);
|
||||||
}
|
}
|
||||||
|
|
||||||
static const struct fw_card_driver ohci_driver = {
|
static const struct fw_card_driver ohci_driver = {
|
||||||
|
@ -425,7 +425,8 @@ struct fw_card_driver {
|
|||||||
int node_id, int generation);
|
int node_id, int generation);
|
||||||
|
|
||||||
struct fw_iso_context *
|
struct fw_iso_context *
|
||||||
(*allocate_iso_context)(struct fw_card *card, int type);
|
(*allocate_iso_context)(struct fw_card *card,
|
||||||
|
int type, size_t header_size);
|
||||||
void (*free_iso_context)(struct fw_iso_context *ctx);
|
void (*free_iso_context)(struct fw_iso_context *ctx);
|
||||||
|
|
||||||
int (*start_iso)(struct fw_iso_context *ctx, s32 cycle);
|
int (*start_iso)(struct fw_iso_context *ctx, s32 cycle);
|
||||||
|
Loading…
Reference in New Issue
Block a user