[kernel] r10007 - in dists/trunk/linux-2.6/debian: . patches/bugfix/all patches/series

Maximilian Attems maks at alioth.debian.org
Mon Dec 24 00:39:41 UTC 2007


Author: maks
Date: Mon Dec 24 00:39:41 2007
New Revision: 10007

Log:
add latest git firewire fixes


Added:
   dists/trunk/linux-2.6/debian/patches/bugfix/all/git-ieee1394.patch
Modified:
   dists/trunk/linux-2.6/debian/changelog
   dists/trunk/linux-2.6/debian/patches/series/1~experimental.1

Modified: dists/trunk/linux-2.6/debian/changelog
==============================================================================
--- dists/trunk/linux-2.6/debian/changelog	(original)
+++ dists/trunk/linux-2.6/debian/changelog	Mon Dec 24 00:39:41 2007
@@ -26,6 +26,7 @@
   * Reenable DABUSB as firmware is BSD licensed.
   * [hppa]: Disable OCFS2, due build trouble.
   * topconfig: Enable delay accounting TASKSTATS. (closes: #433204)
+  * Add git-ieee1394.patch for latest firewire fixes.
 
   [ Bastian Blank ]
   * [amd64, i386]: Set kernel architecture to x86.
@@ -40,7 +41,7 @@
   [ dann frazier ]
   * [ia64]: Enable BLK_CPQ_DA
 
- -- maximilian attems <maks at debian.org>  Tue, 18 Dec 2007 22:50:02 +0100
+ -- maximilian attems <maks at debian.org>  Mon, 24 Dec 2007 01:31:44 +0100
 
 linux-2.6 (2.6.23-1~experimental.1) UNRELEASED; urgency=low
 

Added: dists/trunk/linux-2.6/debian/patches/bugfix/all/git-ieee1394.patch
==============================================================================
--- (empty file)
+++ dists/trunk/linux-2.6/debian/patches/bugfix/all/git-ieee1394.patch	Mon Dec 24 00:39:41 2007
@@ -0,0 +1,1133 @@
+GIT 948b907d3fb2d02f77aea8d5b015cc533bff8d65 git+ssh://master.kernel.org/pub/scm/linux/kernel/git/ieee1394/linux1394-2.6.git
+
+commit 
+Author: Stefan Richter <stefanr at s5r6.in-berlin.de>
+Date:   Sat Dec 22 22:14:52 2007 +0100
+
+    firewire: fw-ohci: CycleTooLong interrupt management
+    
+    The firewire-ohci driver so far lacked the ability to resume cycle
+    master duty after that condition happened, as added to ohci1394 in Linux
+    2.6.18 by commit 57fdb58fa5a140bdd52cf4c4ffc30df73676f0a5.  This ports
+    this patch to fw-ohci.
+    
+    The "cycle too long" condition has been seen in practice
+      - with IIDC cameras if a mode with packets too large for a speed is
+        chosen,
+      - sporadically when capturing DV on a VIA VT6306 card with ohci1394/
+        ieee1394/ raw1394/ dvgrab 2.
+        https://bugzilla.redhat.com/show_bug.cgi?id=415841#c7
+    (This does not fix Fedora bug 415841.)
+    
+    Signed-off-by: Stefan Richter <stefanr at s5r6.in-berlin.de>
+
+commit f22532d403d2b91ec355c3dc008688bb87013600
+Author: Rabin Vincent <rabin at rab.in>
+Date:   Fri Dec 21 23:02:15 2007 +0530
+
+    firewire: Fix extraction of source node id
+    
+    Fix extraction of the source node id from the packet header.
+    
+    Signed-off-by: Rabin Vincent <rabin at rab.in>
+    Signed-off-by: Stefan Richter <stefanr at s5r6.in-berlin.de>
+
+commit 47349df008fd1ebcf6512200e9e2eccb313952c4
+Author: David Moore <dcm at MIT.EDU>
+Date:   Wed Dec 19 15:26:38 2007 -0500
+
+    firewire: fw-ohci: Bug fixes for packet-per-buffer support
+    
+    This patch corrects a number of bugs in the current OHCI 1.0
+    packet-per-buffer support:
+    
+    1. Correctly deal with payloads that cross a page boundary.  The
+    previous version would not split the descriptor at such a boundary,
+    potentially corrupting unrelated memory.
+    
+    2. Allow user-space to specify multiple packets per struct
+    fw_cdev_iso_packet in the same way that dual-buffer allows.  This is
+    signaled by header_length being a multiple of header_size.  This
+    multiple determines the number of packets.  The payload size allocated
+    per packet is determined by dividing the total payload size by the
+    number of packets.
+    
+    3. Make sync support work properly for packet-per-buffer.
+    
+    I have tested this patch with libdc1394 by forcing my OHCI 1.1
+    controller to use the packet-per-buffer support instead of dual-buffer.
+    
+    I would greatly appreciate testing by those who have a DV devices and
+    other types of iso streamers to make sure I didn't cause any
+    regressions.
+    
+    Stefan, with this patch, I'm hoping that libdc1394 will work with all
+    your OHCI 1.0 controllers now.
+    
+    The one bit of future work that remains for packet-per-buffer support is
+    the automatic compaction of short payloads that I discussed with
+    Kristian.
+    
+    Signed-off-by: David Moore <dcm at acm.org>
+    Signed-off-by: Stefan Richter <stefanr at s5r6.in-berlin.de>
+
+commit e9f5ca46377ac60a6b7d52c6c19a1661c87c6e20
+Author: David Moore <dcm at MIT.EDU>
+Date:   Wed Dec 19 03:09:18 2007 -0500
+
+    firewire: fw-ohci: Fix for dualbuffer three-or-more buffers
+    
+    This patch fixes the problem where different OHCI 1.1 controllers behave
+    differently when a received iso packet straddles three or more buffers
+    when using the dual-buffer receive mode.  Two changes are made in order
+    to handle this situation:
+    
+    1. The packet sync DMA descriptor is given a non-zero header length and
+    non-zero payload length.  This is because zero-payload descriptors are
+    not discussed in the OHCI 1.1 specs and their behavior is thus
+    undefined.  Instead we use a header size just large enough for a single
+    header and a payload length of 4 bytes for this first descriptor.
+    
+    2. As we process received packets in the context's tasklet, read the
+    packet length out of the headers.  Keep track of the running total of
+    the packet length as "excess_bytes", so we can ignore any descriptors
+    where no packet starts or ends.  These descriptors may not have had
+    their first_res_count or second_res_count fields updated by the
+    controller so we cannot rely on those values.
+    
+    The main drawback of this patch is that the excess_bytes value might get
+    "out of sync" with the packet descriptors if something strange happens
+    to the DMA program.  I'm not if such a thing could ever happen, but I
+    appreciate any suggestions in making it more robust.
+    
+    Also, the packet-per-buffer support may need a similar fix to deal with
+    issue 1, but I haven't done any work on that yet.
+    
+    Stefan, I'm hoping that with this patch, all your OHCI 1.1 controllers
+    will work properly with an unmodified version of libdc1394.
+    
+    Signed-off-by: David Moore <dcm at acm.org>
+    Signed-off-by: Stefan Richter <stefanr at s5r6.in-berlin.de>
+
+commit 070ca2f30c2bbaeeeb740dfad01cc9a27905e6a9
+Author: Stefan Richter <stefanr at s5r6.in-berlin.de>
+Date:   Sun Dec 16 20:53:13 2007 +0100
+
+    ieee1394: ohci1394: don't schedule IT tasklets on IR events
+    
+    Bug noted by Pieter Palmers:  Isochronous transmit tasklets were
+    scheduled on isochronous receive events, in addition to the proper
+    isochronous receive tasklets.
+    
+    http://marc.info/?l=linux1394-devel&m=119783196222802
+    
+    Signed-off-by: Stefan Richter <stefanr at s5r6.in-berlin.de>
+
+commit 27aa95c9e41622c0d4f5c8d30b62abae0cc9ada0
+Author: Stefan Richter <stefanr at s5r6.in-berlin.de>
+Date:   Sun Dec 16 17:32:11 2007 +0100
+
+    firewire: fw-sbp2: remove unused misleading macro
+    
+    SBP2_MAX_SECTORS is nowhere used in fw-sbp2.
+    It merely got copied over from sbp2 where it played a role in the past.
+    
+    Signed-off-by: Stefan Richter <stefanr at s5r6.in-berlin.de>
+
+commit 332319e2852838221b2ece1389248414e060cc94
+Author: Stefan Richter <stefanr at s5r6.in-berlin.de>
+Date:   Sun Dec 16 17:31:26 2007 +0100
+
+    ieee1394: sbp2: raise default transfer size limit
+    
+    This patch speeds up sbp2 a little bit --- but more importantly, it
+    brings the behavior of sbp2 and fw-sbp2 closer to each other.  Like
+    fw-sbp2, sbp2 now does not limit the size of single transfers to 255
+    sectors anymore, unless told so by a blacklist flag or by module load
+    parameters.
+    
+    Only very old bridge chips have been known to need the 255 sectors
+    limit, and we have got one such chip in our hardwired blacklist.  There
+    certainly is a danger that more bridges need that limit; but I prefer to
+    have this issue present in both fw-sbp2 and sbp2 rather than just one of
+    them.
+    
+    An OXUF922 with 400GB 7200RPM disk on an S400 controller is sped up by
+    this patch from 22.9 to 23.5 MB/s according to hdparm.  The same effect
+    could be achieved before by setting a higher max_sectors module
+    parameter.  On buses which use 1394b beta mode, sbp2 and fw-sbp2 will
+    now achieve virtually the same bandwidth.  Fw-sbp2 only remains faster
+    on 1394a buses due to fw-core's gap count optimization.
+    
+    Signed-off-by: Stefan Richter <stefanr at s5r6.in-berlin.de>
+
+commit bc2dfcc923803ab9a60e5316748e25d425a2bd08
+Author: Stefan Richter <stefanr at s5r6.in-berlin.de>
+Date:   Sat Dec 15 14:11:41 2007 +0100
+
+    ieee1394: remove unused code
+    
+    The code has been in "#if 0 - #endif" since Linux 2.6.12.
+    
+    Signed-off-by: Stefan Richter <stefanr at s5r6.in-berlin.de>
+
+commit aa541d501d5be17ba05e8e6374371c5b376ab994
+Author: Stefan Richter <stefanr at s5r6.in-berlin.de>
+Date:   Sat Dec 15 14:04:42 2007 +0100
+
+    ieee1394: small cleanup after "nopage"
+    
+    Signed-off-by: Stefan Richter <stefanr at s5r6.in-berlin.de>
+
+commit 598c25878bf4e7de677079022c42635ebd846e62
+Author: Stefan Richter <stefanr at s5r6.in-berlin.de>
+Date:   Sat Dec 22 21:53:33 2007 +0100
+
+    Revert "firewire: fw-ohci: CycleTooLong interrupt management"
+    
+    This reverts commit 5bd0c4ef883f5e3c24ab91127de0292ebd0fa405.
+    Needs to be updated to rate-limit the kernel log message.
+
+commit 0394d46dc8485840992f6dd57e39f2336e85a6fe
+Author: Nick Piggin <npiggin at suse.de>
+Date:   Wed Dec 5 18:15:53 2007 +1100
+
+    ieee1394: nopage
+    
+    Convert ieee1394 from nopage to fault.
+    Remove redundant vma range checks (correct resource range check is retained).
+    
+    Signed-off-by: Nick Piggin <npiggin at suse.de>
+    Signed-off-by: Andrew Morton <akpm at linux-foundation.org>
+    Signed-off-by: Stefan Richter <stefanr at s5r6.in-berlin.de>
+
+commit 5bd0c4ef883f5e3c24ab91127de0292ebd0fa405
+Author: Stefan Richter <stefanr at s5r6.in-berlin.de>
+Date:   Sun Dec 9 14:03:09 2007 +0100
+
+    firewire: fw-ohci: CycleTooLong interrupt management
+    
+    According to a report by Robin Theander, VIA VT6306 may sporadically
+    trip the "isochronous cycle too long" condition when capturing DV in
+    buffer-fill mode with ohci1394/ ieee1394/ raw1394/ dvgrab 2.
+    https://bugzilla.redhat.com/show_bug.cgi?id=415841#c7
+    
+    The firewire-ohci driver so far lacked the ability to resume cycle
+    master duty after that condition happened, an ability added to ohci1394
+    in Linux 2.6.18 by a patch from Jean-Baptiste Mur (commit
+    57fdb58fa5a140bdd52cf4c4ffc30df73676f0a5).  This ports this commit to
+    fw-ohci just to be sure, since this condition can potentially also
+    happen with fw-ohci.
+    
+    This does alas not fix above referenced Fedora bug 415841.
+    
+    Signed-off-by: Stefan Richter <stefanr at s5r6.in-berlin.de>
+
+commit 1526cb4169cce7b87db54c47ce0fd0c1bd7fb16a
+Author: Joe Perches <joe at perches.com>
+Date:   Mon Nov 19 17:48:10 2007 -0800
+
+    ieee1394: Add missing "space"
+    
+    Signed-off-by: Joe Perches <joe at perches.com>
+    Signed-off-by: Stefan Richter <stefanr at s5r6.in-berlin.de>
+
+commit b4be6170ef896e3a98a71e03f3514ccff264ffde
+Author: Jay Fenlason <fenlason at redhat.com>
+Date:   Wed Nov 7 17:39:00 2007 -0500
+
+    firewire: fw-sbp2: quiet logout errors on device removal
+    
+    This suppresses both reply timed out and management write failed
+    errors on LOGOUT requests.
+    
+    Signed-off-by: Jay Fenlason <fenlason at redhat.com>
+    Signed-off-by: Stefan Richter <stefanr at s5r6.in-berlin.de>
+
+commit 8f50ff61ed0282179371cbef173b8b0aad0d1313
+Author: Stefan Richter <stefanr at s5r6.in-berlin.de>
+Date:   Sun Nov 4 14:59:24 2007 +0100
+
+    ieee1394: sbp2: s/g list access cosmetics
+    
+    Replace sg->length by sg_dma_len(sg).  Rename a variable for shorter
+    line lengths and eliminate some superfluous local variables.
+    
+    Signed-off-by: Stefan Richter <stefanr at s5r6.in-berlin.de>
+
+commit fe702f621c6bdead79dd4172cd00b35ece4b88c3
+Author: Stefan Richter <stefanr at s5r6.in-berlin.de>
+Date:   Sun Nov 4 14:58:43 2007 +0100
+
+    ieee1394: sbp2: enable s/g chaining
+    
+    Signed-off-by: Stefan Richter <stefanr at s5r6.in-berlin.de>
+
+commit 53e11c39606617de4fea57077891abb3870ff383
+Author: Stefan Richter <stefanr at s5r6.in-berlin.de>
+Date:   Sun Nov 4 14:58:11 2007 +0100
+
+    firewire: fw-sbp2: enable s/g chaining
+    
+    Signed-off-by: Stefan Richter <stefanr at s5r6.in-berlin.de>
+
+commit c3e0d276f016e052dfd87b73041a5be6dd08454d
+Author: Stefan Richter <stefanr at s5r6.in-berlin.de>
+Date:   Wed Nov 7 01:12:51 2007 +0100
+
+    firewire: fw-sbp2: refactor workq and kref handling
+    
+    This somewhat reduces the size of firewire-sbp2.ko.
+    
+    Signed-off-by: Stefan Richter <stefanr at s5r6.in-berlin.de>
+ drivers/firewire/fw-ohci.c               |  154 ++++++++++++++++--------------
+ drivers/firewire/fw-sbp2.c               |   85 ++++++++++-------
+ drivers/firewire/fw-transaction.c        |    2 +-
+ drivers/ieee1394/dma.c                   |   39 +++-----
+ drivers/ieee1394/ieee1394_transactions.c |   68 -------------
+ drivers/ieee1394/ohci1394.c              |   12 ++-
+ drivers/ieee1394/raw1394.c               |    4 +-
+ drivers/ieee1394/sbp2.c                  |   53 ++++++-----
+ drivers/ieee1394/sbp2.h                  |    1 -
+ 9 files changed, 184 insertions(+), 234 deletions(-)
+
+diff --git a/drivers/firewire/fw-ohci.c b/drivers/firewire/fw-ohci.c
+index 436a855..74d5d94 100644
+--- a/drivers/firewire/fw-ohci.c
++++ b/drivers/firewire/fw-ohci.c
+@@ -125,6 +125,7 @@ struct context {
+ struct iso_context {
+ 	struct fw_iso_context base;
+ 	struct context context;
++	int excess_bytes;
+ 	void *header;
+ 	size_t header_length;
+ };
+@@ -1078,6 +1079,13 @@ static irqreturn_t irq_handler(int irq, void *data)
+ 	if (unlikely(event & OHCI1394_postedWriteErr))
+ 		fw_error("PCI posted write error\n");
+ 
++	if (unlikely(event & OHCI1394_cycleTooLong)) {
++		if (printk_ratelimit())
++			fw_notify("isochronous cycle too long\n");
++		reg_write(ohci, OHCI1394_LinkControlSet,
++			  OHCI1394_LinkControl_cycleMaster);
++	}
++
+ 	if (event & OHCI1394_cycle64Seconds) {
+ 		cycle_time = reg_read(ohci, OHCI1394_IsochronousCycleTimer);
+ 		if ((cycle_time & 0x80000000) == 0)
+@@ -1151,8 +1159,8 @@ static int ohci_enable(struct fw_card *card, u32 *config_rom, size_t length)
+ 		  OHCI1394_RQPkt | OHCI1394_RSPkt |
+ 		  OHCI1394_reqTxComplete | OHCI1394_respTxComplete |
+ 		  OHCI1394_isochRx | OHCI1394_isochTx |
+-		  OHCI1394_postedWriteErr | OHCI1394_cycle64Seconds |
+-		  OHCI1394_masterIntEnable);
++		  OHCI1394_postedWriteErr | OHCI1394_cycleTooLong |
++		  OHCI1394_cycle64Seconds | OHCI1394_masterIntEnable);
+ 
+ 	/* Activate link_on bit and contender bit in our self ID packets.*/
+ 	if (ohci_update_phy_reg(card, 4, 0,
+@@ -1408,9 +1416,13 @@ static int handle_ir_dualbuffer_packet(struct context *context,
+ 	void *p, *end;
+ 	int i;
+ 
+-	if (db->first_res_count > 0 && db->second_res_count > 0)
+-		/* This descriptor isn't done yet, stop iteration. */
+-		return 0;
++	if (db->first_res_count > 0 && db->second_res_count > 0) {
++		if (ctx->excess_bytes <= le16_to_cpu(db->second_req_count)) {
++			/* This descriptor isn't done yet, stop iteration. */
++			return 0;
++		}
++		ctx->excess_bytes -= le16_to_cpu(db->second_req_count);
++	}
+ 
+ 	header_length = le16_to_cpu(db->first_req_count) -
+ 		le16_to_cpu(db->first_res_count);
+@@ -1429,11 +1441,15 @@ static int handle_ir_dualbuffer_packet(struct context *context,
+ 		*(u32 *) (ctx->header + i) = __swab32(*(u32 *) (p + 4));
+ 		memcpy(ctx->header + i + 4, p + 8, ctx->base.header_size - 4);
+ 		i += ctx->base.header_size;
++		ctx->excess_bytes +=
++			(le32_to_cpu(*(u32 *)(p + 4)) >> 16) & 0xffff;
+ 		p += ctx->base.header_size + 4;
+ 	}
+-
+ 	ctx->header_length = i;
+ 
++	ctx->excess_bytes -= le16_to_cpu(db->second_req_count) -
++		le16_to_cpu(db->second_res_count);
++
+ 	if (le16_to_cpu(db->control) & DESCRIPTOR_IRQ_ALWAYS) {
+ 		ir_header = (__le32 *) (db + 1);
+ 		ctx->base.callback(&ctx->base,
+@@ -1452,24 +1468,24 @@ static int handle_ir_packet_per_buffer(struct context *context,
+ {
+ 	struct iso_context *ctx =
+ 		container_of(context, struct iso_context, context);
+-	struct descriptor *pd = d + 1;
++	struct descriptor *pd;
+ 	__le32 *ir_header;
+-	size_t header_length;
+-	void *p, *end;
+-	int i, z;
++	void *p;
++	int i;
+ 
+-	if (pd->res_count == pd->req_count)
++	for (pd = d; pd <= last; pd++) {
++		if (pd->transfer_status)
++			break;
++	}
++	if (pd > last)
+ 		/* Descriptor(s) not done yet, stop iteration */
+ 		return 0;
+ 
+-	header_length = le16_to_cpu(d->req_count);
+-
+ 	i   = ctx->header_length;
+-	z   = le32_to_cpu(pd->branch_address) & 0xf;
+-	p   = d + z;
+-	end = p + header_length;
++	p   = last + 1;
+ 
+-	while (p < end && i + ctx->base.header_size <= PAGE_SIZE) {
++	if (ctx->base.header_size > 0 &&
++			i + ctx->base.header_size <= PAGE_SIZE) {
+ 		/*
+ 		 * The iso header is byteswapped to little endian by
+ 		 * the controller, but the remaining header quadlets
+@@ -1478,14 +1494,11 @@ static int handle_ir_packet_per_buffer(struct context *context,
+ 		 */
+ 		*(u32 *) (ctx->header + i) = __swab32(*(u32 *) (p + 4));
+ 		memcpy(ctx->header + i + 4, p + 8, ctx->base.header_size - 4);
+-		i += ctx->base.header_size;
+-		p += ctx->base.header_size + 4;
++		ctx->header_length += ctx->base.header_size;
+ 	}
+ 
+-	ctx->header_length = i;
+-
+-	if (le16_to_cpu(pd->control) & DESCRIPTOR_IRQ_ALWAYS) {
+-		ir_header = (__le32 *) (d + z);
++	if (le16_to_cpu(last->control) & DESCRIPTOR_IRQ_ALWAYS) {
++		ir_header = (__le32 *) p;
+ 		ctx->base.callback(&ctx->base,
+ 				   le32_to_cpu(ir_header[0]) & 0xffff,
+ 				   ctx->header_length, ctx->header,
+@@ -1493,7 +1506,6 @@ static int handle_ir_packet_per_buffer(struct context *context,
+ 		ctx->header_length = 0;
+ 	}
+ 
+-
+ 	return 1;
+ }
+ 
+@@ -1775,19 +1787,6 @@ ohci_queue_iso_receive_dualbuffer(struct fw_iso_context *base,
+ 	 * packet, retransmit or terminate..
+ 	 */
+ 
+-	if (packet->skip) {
+-		d = context_get_descriptors(&ctx->context, 2, &d_bus);
+-		if (d == NULL)
+-			return -ENOMEM;
+-
+-		db = (struct db_descriptor *) d;
+-		db->control = cpu_to_le16(DESCRIPTOR_STATUS |
+-					  DESCRIPTOR_BRANCH_ALWAYS |
+-					  DESCRIPTOR_WAIT);
+-		db->first_size = cpu_to_le16(ctx->base.header_size + 4);
+-		context_append(&ctx->context, d, 2, 0);
+-	}
+-
+ 	p = packet;
+ 	z = 2;
+ 
+@@ -1815,11 +1814,18 @@ ohci_queue_iso_receive_dualbuffer(struct fw_iso_context *base,
+ 		db->control = cpu_to_le16(DESCRIPTOR_STATUS |
+ 					  DESCRIPTOR_BRANCH_ALWAYS);
+ 		db->first_size = cpu_to_le16(ctx->base.header_size + 4);
+-		db->first_req_count = cpu_to_le16(header_size);
++		if (p->skip && rest == p->payload_length) {
++			db->control |= cpu_to_le16(DESCRIPTOR_WAIT);
++			db->first_req_count = db->first_size;
++		} else {
++			db->first_req_count = cpu_to_le16(header_size);
++		}
+ 		db->first_res_count = db->first_req_count;
+ 		db->first_buffer = cpu_to_le32(d_bus + sizeof(*db));
+ 
+-		if (offset + rest < PAGE_SIZE)
++		if (p->skip && rest == p->payload_length)
++			length = 4;
++		else if (offset + rest < PAGE_SIZE)
+ 			length = rest;
+ 		else
+ 			length = PAGE_SIZE - offset;
+@@ -1835,7 +1841,8 @@ ohci_queue_iso_receive_dualbuffer(struct fw_iso_context *base,
+ 		context_append(&ctx->context, d, z, header_z);
+ 		offset = (offset + length) & ~PAGE_MASK;
+ 		rest -= length;
+-		page++;
++		if (offset == 0)
++			page++;
+ 	}
+ 
+ 	return 0;
+@@ -1849,67 +1856,70 @@ ohci_queue_iso_receive_packet_per_buffer(struct fw_iso_context *base,
+ {
+ 	struct iso_context *ctx = container_of(base, struct iso_context, base);
+ 	struct descriptor *d = NULL, *pd = NULL;
+-	struct fw_iso_packet *p;
++	struct fw_iso_packet *p = packet;
+ 	dma_addr_t d_bus, page_bus;
+ 	u32 z, header_z, rest;
+-	int i, page, offset, packet_count, header_size;
+-
+-	if (packet->skip) {
+-		d = context_get_descriptors(&ctx->context, 1, &d_bus);
+-		if (d == NULL)
+-			return -ENOMEM;
+-
+-		d->control = cpu_to_le16(DESCRIPTOR_STATUS |
+-					 DESCRIPTOR_INPUT_LAST |
+-					 DESCRIPTOR_BRANCH_ALWAYS |
+-					 DESCRIPTOR_WAIT);
+-		context_append(&ctx->context, d, 1, 0);
+-	}
+-
+-	/* one descriptor for header, one for payload */
+-	/* FIXME: handle cases where we need multiple desc. for payload */
+-	z = 2;
+-	p = packet;
++	int i, j, length;
++	int page, offset, packet_count, header_size, payload_per_buffer;
+ 
+ 	/*
+ 	 * The OHCI controller puts the status word in the
+ 	 * buffer too, so we need 4 extra bytes per packet.
+ 	 */
+ 	packet_count = p->header_length / ctx->base.header_size;
+-	header_size  = packet_count * (ctx->base.header_size + 4);
++	header_size  = ctx->base.header_size + 4;
+ 
+ 	/* Get header size in number of descriptors. */
+ 	header_z = DIV_ROUND_UP(header_size, sizeof(*d));
+ 	page     = payload >> PAGE_SHIFT;
+ 	offset   = payload & ~PAGE_MASK;
+-	rest     = p->payload_length;
++	payload_per_buffer = p->payload_length / packet_count;
+ 
+ 	for (i = 0; i < packet_count; i++) {
+ 		/* d points to the header descriptor */
++		z = DIV_ROUND_UP(payload_per_buffer + offset, PAGE_SIZE) + 1;
+ 		d = context_get_descriptors(&ctx->context,
+-					    z + header_z, &d_bus);
++				z + header_z, &d_bus);
+ 		if (d == NULL)
+ 			return -ENOMEM;
+ 
+-		d->control      = cpu_to_le16(DESCRIPTOR_INPUT_MORE);
++		d->control      = cpu_to_le16(DESCRIPTOR_STATUS |
++					      DESCRIPTOR_INPUT_MORE);
++		if (p->skip && i == 0)
++			d->control |= cpu_to_le16(DESCRIPTOR_WAIT);
+ 		d->req_count    = cpu_to_le16(header_size);
+ 		d->res_count    = d->req_count;
++		d->transfer_status = 0;
+ 		d->data_address = cpu_to_le32(d_bus + (z * sizeof(*d)));
+ 
+-		/* pd points to the payload descriptor */
+-		pd = d + 1;
++		rest = payload_per_buffer;
++		for (j = 1; j < z; j++) {
++			pd = d + j;
++			pd->control = cpu_to_le16(DESCRIPTOR_STATUS |
++						  DESCRIPTOR_INPUT_MORE);
++
++			if (offset + rest < PAGE_SIZE)
++				length = rest;
++			else
++				length = PAGE_SIZE - offset;
++			pd->req_count = cpu_to_le16(length);
++			pd->res_count = pd->req_count;
++			pd->transfer_status = 0;
++
++			page_bus = page_private(buffer->pages[page]);
++			pd->data_address = cpu_to_le32(page_bus + offset);
++
++			offset = (offset + length) & ~PAGE_MASK;
++			rest -= length;
++			if (offset == 0)
++				page++;
++		}
+ 		pd->control = cpu_to_le16(DESCRIPTOR_STATUS |
+ 					  DESCRIPTOR_INPUT_LAST |
+ 					  DESCRIPTOR_BRANCH_ALWAYS);
+-		if (p->interrupt)
++		if (p->interrupt && i == packet_count - 1)
+ 			pd->control |= cpu_to_le16(DESCRIPTOR_IRQ_ALWAYS);
+ 
+-		pd->req_count = cpu_to_le16(rest);
+-		pd->res_count = pd->req_count;
+-
+-		page_bus = page_private(buffer->pages[page]);
+-		pd->data_address = cpu_to_le32(page_bus + offset);
+-
+ 		context_append(&ctx->context, d, z, header_z);
+ 	}
+ 
+diff --git a/drivers/firewire/fw-sbp2.c b/drivers/firewire/fw-sbp2.c
+index 624ff3e..9040417 100644
+--- a/drivers/firewire/fw-sbp2.c
++++ b/drivers/firewire/fw-sbp2.c
+@@ -151,9 +151,7 @@ struct sbp2_target {
+ };
+ 
+ #define SBP2_MAX_SG_ELEMENT_LENGTH	0xf000
+-#define SBP2_MAX_SECTORS		255	/* Max sectors supported */
+ #define SBP2_ORB_TIMEOUT		2000	/* Timeout in ms */
+-
+ #define SBP2_ORB_NULL			0x80000000
+ 
+ #define SBP2_DIRECTION_TO_MEDIA		0x0
+@@ -540,14 +538,26 @@ sbp2_send_management_orb(struct sbp2_logical_unit *lu, int node_id,
+ 
+ 	retval = -EIO;
+ 	if (sbp2_cancel_orbs(lu) == 0) {
+-		fw_error("orb reply timed out, rcode=0x%02x\n",
+-			 orb->base.rcode);
++		/*
++		 * Logout requests frequently get sent to devices that aren't
++		 * there any more, resulting in extraneous error messages in
++		 * the logs.  Unfortunately, this means logout requests that
++		 * actually fail don't get logged.
++		 */
++		if (function != SBP2_LOGOUT_REQUEST)
++			fw_error("orb reply timed out, rcode=0x%02x\n",
++				 orb->base.rcode);
+ 		goto out;
+ 	}
+ 
+ 	if (orb->base.rcode != RCODE_COMPLETE) {
+-		fw_error("management write failed, rcode 0x%02x\n",
+-			 orb->base.rcode);
++		/*
++		 * On device removal from the bus, sometimes the logout
++		 * request times out, sometimes it just fails.
++		 */
++		if (function != SBP2_LOGOUT_REQUEST)
++			fw_error("management write failed, rcode 0x%02x\n",
++				 orb->base.rcode);
+ 		goto out;
+ 	}
+ 
+@@ -628,6 +638,21 @@ static void sbp2_release_target(struct kref *kref)
+ 
+ static struct workqueue_struct *sbp2_wq;
+ 
++/*
++ * Always get the target's kref when scheduling work on one its units.
++ * Each workqueue job is responsible to call sbp2_target_put() upon return.
++ */
++static void sbp2_queue_work(struct sbp2_logical_unit *lu, unsigned long delay)
++{
++	if (queue_delayed_work(sbp2_wq, &lu->work, delay))
++		kref_get(&lu->tgt->kref);
++}
++
++static void sbp2_target_put(struct sbp2_target *tgt)
++{
++	kref_put(&tgt->kref, sbp2_release_target);
++}
++
+ static void sbp2_reconnect(struct work_struct *work);
+ 
+ static void sbp2_login(struct work_struct *work)
+@@ -649,16 +674,12 @@ static void sbp2_login(struct work_struct *work)
+ 
+ 	if (sbp2_send_management_orb(lu, node_id, generation,
+ 				SBP2_LOGIN_REQUEST, lu->lun, &response) < 0) {
+-		if (lu->retries++ < 5) {
+-			if (queue_delayed_work(sbp2_wq, &lu->work,
+-					       DIV_ROUND_UP(HZ, 5)))
+-				kref_get(&lu->tgt->kref);
+-		} else {
++		if (lu->retries++ < 5)
++			sbp2_queue_work(lu, DIV_ROUND_UP(HZ, 5));
++		else
+ 			fw_error("failed to login to %s LUN %04x\n",
+ 				 unit->device.bus_id, lu->lun);
+-		}
+-		kref_put(&lu->tgt->kref, sbp2_release_target);
+-		return;
++		goto out;
+ 	}
+ 
+ 	lu->generation        = generation;
+@@ -700,7 +721,8 @@ static void sbp2_login(struct work_struct *work)
+ 		lu->sdev = sdev;
+ 		scsi_device_put(sdev);
+ 	}
+-	kref_put(&lu->tgt->kref, sbp2_release_target);
++ out:
++	sbp2_target_put(lu->tgt);
+ }
+ 
+ static int sbp2_add_logical_unit(struct sbp2_target *tgt, int lun_entry)
+@@ -865,18 +887,13 @@ static int sbp2_probe(struct device *dev)
+ 
+ 	get_device(&unit->device);
+ 
+-	/*
+-	 * We schedule work to do the login so we can easily
+-	 * reschedule retries. Always get the ref before scheduling
+-	 * work.
+-	 */
++	/* Do the login in a workqueue so we can easily reschedule retries. */
+ 	list_for_each_entry(lu, &tgt->lu_list, link)
+-		if (queue_delayed_work(sbp2_wq, &lu->work, 0))
+-			kref_get(&tgt->kref);
++		sbp2_queue_work(lu, 0);
+ 	return 0;
+ 
+  fail_tgt_put:
+-	kref_put(&tgt->kref, sbp2_release_target);
++	sbp2_target_put(tgt);
+ 	return -ENOMEM;
+ 
+  fail_shost_put:
+@@ -889,7 +906,7 @@ static int sbp2_remove(struct device *dev)
+ 	struct fw_unit *unit = fw_unit(dev);
+ 	struct sbp2_target *tgt = unit->device.driver_data;
+ 
+-	kref_put(&tgt->kref, sbp2_release_target);
++	sbp2_target_put(tgt);
+ 	return 0;
+ }
+ 
+@@ -915,10 +932,8 @@ static void sbp2_reconnect(struct work_struct *work)
+ 			lu->retries = 0;
+ 			PREPARE_DELAYED_WORK(&lu->work, sbp2_login);
+ 		}
+-		if (queue_delayed_work(sbp2_wq, &lu->work, DIV_ROUND_UP(HZ, 5)))
+-			kref_get(&lu->tgt->kref);
+-		kref_put(&lu->tgt->kref, sbp2_release_target);
+-		return;
++		sbp2_queue_work(lu, DIV_ROUND_UP(HZ, 5));
++		goto out;
+ 	}
+ 
+ 	lu->generation        = generation;
+@@ -930,8 +945,8 @@ static void sbp2_reconnect(struct work_struct *work)
+ 
+ 	sbp2_agent_reset(lu);
+ 	sbp2_cancel_orbs(lu);
+-
+-	kref_put(&lu->tgt->kref, sbp2_release_target);
++ out:
++	sbp2_target_put(lu->tgt);
+ }
+ 
+ static void sbp2_update(struct fw_unit *unit)
+@@ -947,8 +962,7 @@ static void sbp2_update(struct fw_unit *unit)
+ 	 */
+ 	list_for_each_entry(lu, &tgt->lu_list, link) {
+ 		lu->retries = 0;
+-		if (queue_delayed_work(sbp2_wq, &lu->work, 0))
+-			kref_get(&tgt->kref);
++		sbp2_queue_work(lu, 0);
+ 	}
+ }
+ 
+@@ -1103,9 +1117,9 @@ sbp2_map_scatterlist(struct sbp2_command_orb *orb, struct fw_device *device,
+ 	 * elements larger than 65535 bytes, some IOMMUs may merge sg elements
+ 	 * during DMA mapping, and Linux currently doesn't prevent this.
+ 	 */
+-	for (i = 0, j = 0; i < count; i++) {
+-		sg_len = sg_dma_len(sg + i);
+-		sg_addr = sg_dma_address(sg + i);
++	for (i = 0, j = 0; i < count; i++, sg = sg_next(sg)) {
++		sg_len = sg_dma_len(sg);
++		sg_addr = sg_dma_address(sg);
+ 		while (sg_len) {
+ 			/* FIXME: This won't get us out of the pinch. */
+ 			if (unlikely(j >= ARRAY_SIZE(orb->page_table))) {
+@@ -1325,6 +1339,7 @@ static struct scsi_host_template scsi_driver_template = {
+ 	.this_id		= -1,
+ 	.sg_tablesize		= SG_ALL,
+ 	.use_clustering		= ENABLE_CLUSTERING,
++	.use_sg_chaining	= ENABLE_SG_CHAINING,
+ 	.cmd_per_lun		= 1,
+ 	.can_queue		= 1,
+ 	.sdev_attrs		= sbp2_scsi_sysfs_attrs,
+diff --git a/drivers/firewire/fw-transaction.c b/drivers/firewire/fw-transaction.c
+index c00d4a9..8018c3b 100644
+--- a/drivers/firewire/fw-transaction.c
++++ b/drivers/firewire/fw-transaction.c
+@@ -650,7 +650,7 @@ fw_core_handle_request(struct fw_card *card, struct fw_packet *p)
+ 		 HEADER_GET_OFFSET_HIGH(p->header[1]) << 32) | p->header[2];
+ 	tcode       = HEADER_GET_TCODE(p->header[0]);
+ 	destination = HEADER_GET_DESTINATION(p->header[0]);
+-	source      = HEADER_GET_SOURCE(p->header[0]);
++	source      = HEADER_GET_SOURCE(p->header[1]);
+ 
+ 	spin_lock_irqsave(&address_handler_lock, flags);
+ 	handler = lookup_enclosing_address_handler(&address_handler_list,
+diff --git a/drivers/ieee1394/dma.c b/drivers/ieee1394/dma.c
+index 7c4eb39..73685e7 100644
+--- a/drivers/ieee1394/dma.c
++++ b/drivers/ieee1394/dma.c
+@@ -231,37 +231,24 @@ void dma_region_sync_for_device(struct dma_region *dma, unsigned long offset,
+ 
+ #ifdef CONFIG_MMU
+ 
+-/* nopage() handler for mmap access */
+-
+-static struct page *dma_region_pagefault(struct vm_area_struct *area,
+-					 unsigned long address, int *type)
++static int dma_region_pagefault(struct vm_area_struct *vma,
++				struct vm_fault *vmf)
+ {
+-	unsigned long offset;
+-	unsigned long kernel_virt_addr;
+-	struct page *ret = NOPAGE_SIGBUS;
+-
+-	struct dma_region *dma = (struct dma_region *)area->vm_private_data;
++	struct dma_region *dma = (struct dma_region *)vma->vm_private_data;
+ 
+ 	if (!dma->kvirt)
+-		goto out;
+-
+-	if ((address < (unsigned long)area->vm_start) ||
+-	    (address >
+-	     (unsigned long)area->vm_start + (dma->n_pages << PAGE_SHIFT)))
+-		goto out;
+-
+-	if (type)
+-		*type = VM_FAULT_MINOR;
+-	offset = address - area->vm_start;
+-	kernel_virt_addr = (unsigned long)dma->kvirt + offset;
+-	ret = vmalloc_to_page((void *)kernel_virt_addr);
+-	get_page(ret);
+-      out:
+-	return ret;
++		return VM_FAULT_SIGBUS;
++
++	if (vmf->pgoff >= dma->n_pages)
++		return VM_FAULT_SIGBUS;
++
++	vmf->page = vmalloc_to_page(dma->kvirt + (vmf->pgoff << PAGE_SHIFT));
++	get_page(vmf->page);
++	return 0;
+ }
+ 
+ static struct vm_operations_struct dma_region_vm_ops = {
+-	.nopage = dma_region_pagefault,
++	.fault = dma_region_pagefault,
+ };
+ 
+ /**
+@@ -275,7 +262,7 @@ int dma_region_mmap(struct dma_region *dma, struct file *file,
+ 	if (!dma->kvirt)
+ 		return -EINVAL;
+ 
+-	/* must be page-aligned */
++	/* must be page-aligned (XXX: comment is wrong, we could allow pgoff) */
+ 	if (vma->vm_pgoff != 0)
+ 		return -EINVAL;
+ 
+diff --git a/drivers/ieee1394/ieee1394_transactions.c b/drivers/ieee1394/ieee1394_transactions.c
+index 6779893..10c3d9f 100644
+--- a/drivers/ieee1394/ieee1394_transactions.c
++++ b/drivers/ieee1394/ieee1394_transactions.c
+@@ -570,71 +570,3 @@ int hpsb_write(struct hpsb_host *host, nodeid_t node, unsigned int generation,
+ 
+ 	return retval;
+ }
+-
+-#if 0
+-
+-int hpsb_lock(struct hpsb_host *host, nodeid_t node, unsigned int generation,
+-	      u64 addr, int extcode, quadlet_t * data, quadlet_t arg)
+-{
+-	struct hpsb_packet *packet;
+-	int retval = 0;
+-
+-	BUG_ON(in_interrupt());	// We can't be called in an interrupt, yet
+-
+-	packet = hpsb_make_lockpacket(host, node, addr, extcode, data, arg);
+-	if (!packet)
+-		return -ENOMEM;
+-
+-	packet->generation = generation;
+-	retval = hpsb_send_packet_and_wait(packet);
+-	if (retval < 0)
+-		goto hpsb_lock_fail;
+-
+-	retval = hpsb_packet_success(packet);
+-
+-	if (retval == 0) {
+-		*data = packet->data[0];
+-	}
+-
+-      hpsb_lock_fail:
+-	hpsb_free_tlabel(packet);
+-	hpsb_free_packet(packet);
+-
+-	return retval;
+-}
+-
+-int hpsb_send_gasp(struct hpsb_host *host, int channel, unsigned int generation,
+-		   quadlet_t * buffer, size_t length, u32 specifier_id,
+-		   unsigned int version)
+-{
+-	struct hpsb_packet *packet;
+-	int retval = 0;
+-	u16 specifier_id_hi = (specifier_id & 0x00ffff00) >> 8;
+-	u8 specifier_id_lo = specifier_id & 0xff;
+-
+-	HPSB_VERBOSE("Send GASP: channel = %d, length = %Zd", channel, length);
+-
+-	length += 8;
+-
+-	packet = hpsb_make_streampacket(host, NULL, length, channel, 3, 0);
+-	if (!packet)
+-		return -ENOMEM;
+-
+-	packet->data[0] = cpu_to_be32((host->node_id << 16) | specifier_id_hi);
+-	packet->data[1] =
+-	    cpu_to_be32((specifier_id_lo << 24) | (version & 0x00ffffff));
+-
+-	memcpy(&(packet->data[2]), buffer, length - 8);
+-
+-	packet->generation = generation;
+-
+-	packet->no_waiter = 1;
+-
+-	retval = hpsb_send_packet(packet);
+-	if (retval < 0)
+-		hpsb_free_packet(packet);
+-
+-	return retval;
+-}
+-
+-#endif				/*  0  */
+diff --git a/drivers/ieee1394/ohci1394.c b/drivers/ieee1394/ohci1394.c
+index 372c5c1..969de2a 100644
+--- a/drivers/ieee1394/ohci1394.c
++++ b/drivers/ieee1394/ohci1394.c
+@@ -2126,10 +2126,14 @@ static void ohci_schedule_iso_tasklets(struct ti_ohci *ohci,
+ 	list_for_each_entry(t, &ohci->iso_tasklet_list, link) {
+ 		mask = 1 << t->context;
+ 
+-		if (t->type == OHCI_ISO_TRANSMIT && tx_event & mask)
+-			tasklet_schedule(&t->tasklet);
+-		else if (rx_event & mask)
+-			tasklet_schedule(&t->tasklet);
++		if (t->type == OHCI_ISO_TRANSMIT) {
++			if (tx_event & mask)
++				tasklet_schedule(&t->tasklet);
++		} else {
++			/* OHCI_ISO_RECEIVE or OHCI_ISO_MULTICHANNEL_RECEIVE */
++			if (rx_event & mask)
++				tasklet_schedule(&t->tasklet);
++		}
+ 	}
+ 
+ 	spin_unlock_irqrestore(&ohci->iso_tasklet_list_lock, flags);
+diff --git a/drivers/ieee1394/raw1394.c b/drivers/ieee1394/raw1394.c
+index cadf047..37e7e10 100644
+--- a/drivers/ieee1394/raw1394.c
++++ b/drivers/ieee1394/raw1394.c
+@@ -858,7 +858,7 @@ static int arm_read(struct hpsb_host *host, int nodeid, quadlet_t * buffer,
+ 	int found = 0, size = 0, rcode = -1;
+ 	struct arm_request_response *arm_req_resp = NULL;
+ 
+-	DBGMSG("arm_read  called by node: %X"
++	DBGMSG("arm_read  called by node: %X "
+ 	       "addr: %4.4x %8.8x length: %Zu", nodeid,
+ 	       (u16) ((addr >> 32) & 0xFFFF), (u32) (addr & 0xFFFFFFFF),
+ 	       length);
+@@ -1012,7 +1012,7 @@ static int arm_write(struct hpsb_host *host, int nodeid, int destid,
+ 	int found = 0, size = 0, rcode = -1, length_conflict = 0;
+ 	struct arm_request_response *arm_req_resp = NULL;
+ 
+-	DBGMSG("arm_write called by node: %X"
++	DBGMSG("arm_write called by node: %X "
+ 	       "addr: %4.4x %8.8x length: %Zu", nodeid,
+ 	       (u16) ((addr >> 32) & 0xFFFF), (u32) (addr & 0xFFFFFFFF),
+ 	       length);
+diff --git a/drivers/ieee1394/sbp2.c b/drivers/ieee1394/sbp2.c
+index b83d254..d2747f0 100644
+--- a/drivers/ieee1394/sbp2.c
++++ b/drivers/ieee1394/sbp2.c
+@@ -51,6 +51,7 @@
+  * Grep for inline FIXME comments below.
+  */
+ 
++#include <linux/blkdev.h>
+ #include <linux/compiler.h>
+ #include <linux/delay.h>
+ #include <linux/device.h>
+@@ -127,17 +128,21 @@ MODULE_PARM_DESC(serialize_io, "Serialize requests coming from SCSI drivers "
+ 		 "(default = Y, faster but buggy = N)");
+ 
+ /*
+- * Bump up max_sectors if you'd like to support very large sized
+- * transfers. Please note that some older sbp2 bridge chips are broken for
+- * transfers greater or equal to 128KB.  Default is a value of 255
+- * sectors, or just under 128KB (at 512 byte sector size). I can note that
+- * the Oxsemi sbp2 chipsets have no problems supporting very large
+- * transfer sizes.
++ * Adjust max_sectors if you'd like to influence how many sectors each SCSI
++ * command can transfer at most. Please note that some older SBP-2 bridge
++ * chips are broken for transfers greater or equal to 128KB, therefore
++ * max_sectors used to be a safe 255 sectors for many years. We now have a
++ * default of 0 here which means that we let the SCSI stack choose a limit.
++ *
++ * The SBP2_WORKAROUND_128K_MAX_TRANS flag, if set either in the workarounds
++ * module parameter or in the sbp2_workarounds_table[], will override the
++ * value of max_sectors. We should use sbp2_workarounds_table[] to cover any
++ * bridge chip which becomes known to need the 255 sectors limit.
+  */
+-static int sbp2_max_sectors = SBP2_MAX_SECTORS;
++static int sbp2_max_sectors;
+ module_param_named(max_sectors, sbp2_max_sectors, int, 0444);
+ MODULE_PARM_DESC(max_sectors, "Change max sectors per I/O supported "
+-		 "(default = " __stringify(SBP2_MAX_SECTORS) ")");
++		 "(default = 0 = use SCSI stack's default)");
+ 
+ /*
+  * Exclusive login to sbp2 device? In most cases, the sbp2 driver should
+@@ -326,6 +331,7 @@ static struct scsi_host_template sbp2_shost_template = {
+ 	.this_id		 = -1,
+ 	.sg_tablesize		 = SG_ALL,
+ 	.use_clustering		 = ENABLE_CLUSTERING,
++	.use_sg_chaining	 = ENABLE_SG_CHAINING,
+ 	.cmd_per_lun		 = SBP2_MAX_CMDS,
+ 	.can_queue		 = SBP2_MAX_CMDS,
+ 	.sdev_attrs		 = sbp2_sysfs_sdev_attrs,
+@@ -1451,7 +1457,7 @@ static void sbp2_prep_command_orb_sg(struct sbp2_command_orb *orb,
+ 				     struct sbp2_fwhost_info *hi,
+ 				     struct sbp2_command_info *cmd,
+ 				     unsigned int scsi_use_sg,
+-				     struct scatterlist *sgpnt,
++				     struct scatterlist *sg,
+ 				     u32 orb_direction,
+ 				     enum dma_data_direction dma_dir)
+ {
+@@ -1461,12 +1467,12 @@ static void sbp2_prep_command_orb_sg(struct sbp2_command_orb *orb,
+ 
+ 	/* special case if only one element (and less than 64KB in size) */
+ 	if ((scsi_use_sg == 1) &&
+-	    (sgpnt[0].length <= SBP2_MAX_SG_ELEMENT_LENGTH)) {
++	    (sg_dma_len(sg) <= SBP2_MAX_SG_ELEMENT_LENGTH)) {
+ 
+-		cmd->dma_size = sgpnt[0].length;
++		cmd->dma_size = sg_dma_len(sg);
+ 		cmd->dma_type = CMD_DMA_PAGE;
+ 		cmd->cmd_dma = dma_map_page(hi->host->device.parent,
+-					    sg_page(&sgpnt[0]), sgpnt[0].offset,
++					    sg_page(sg), sg->offset,
+ 					    cmd->dma_size, cmd->dma_dir);
+ 
+ 		orb->data_descriptor_lo = cmd->cmd_dma;
+@@ -1477,11 +1483,11 @@ static void sbp2_prep_command_orb_sg(struct sbp2_command_orb *orb,
+ 						&cmd->scatter_gather_element[0];
+ 		u32 sg_count, sg_len;
+ 		dma_addr_t sg_addr;
+-		int i, count = dma_map_sg(hi->host->device.parent, sgpnt,
++		int i, count = dma_map_sg(hi->host->device.parent, sg,
+ 					  scsi_use_sg, dma_dir);
+ 
+ 		cmd->dma_size = scsi_use_sg;
+-		cmd->sge_buffer = sgpnt;
++		cmd->sge_buffer = sg;
+ 
+ 		/* use page tables (s/g) */
+ 		orb->misc |= ORB_SET_PAGE_TABLE_PRESENT(0x1);
+@@ -1489,9 +1495,9 @@ static void sbp2_prep_command_orb_sg(struct sbp2_command_orb *orb,
+ 
+ 		/* loop through and fill out our SBP-2 page tables
+ 		 * (and split up anything too large) */
+-		for (i = 0, sg_count = 0 ; i < count; i++, sgpnt++) {
+-			sg_len = sg_dma_len(sgpnt);
+-			sg_addr = sg_dma_address(sgpnt);
++		for (i = 0, sg_count = 0; i < count; i++, sg = sg_next(sg)) {
++			sg_len = sg_dma_len(sg);
++			sg_addr = sg_dma_address(sg);
+ 			while (sg_len) {
+ 				sg_element[sg_count].segment_base_lo = sg_addr;
+ 				if (sg_len > SBP2_MAX_SG_ELEMENT_LENGTH) {
+@@ -1521,11 +1527,10 @@ static void sbp2_create_command_orb(struct sbp2_lu *lu,
+ 				    unchar *scsi_cmd,
+ 				    unsigned int scsi_use_sg,
+ 				    unsigned int scsi_request_bufflen,
+-				    void *scsi_request_buffer,
++				    struct scatterlist *sg,
+ 				    enum dma_data_direction dma_dir)
+ {
+ 	struct sbp2_fwhost_info *hi = lu->hi;
+-	struct scatterlist *sgpnt = (struct scatterlist *)scsi_request_buffer;
+ 	struct sbp2_command_orb *orb = &cmd->command_orb;
+ 	u32 orb_direction;
+ 
+@@ -1560,7 +1565,7 @@ static void sbp2_create_command_orb(struct sbp2_lu *lu,
+ 		orb->data_descriptor_lo = 0x0;
+ 		orb->misc |= ORB_SET_DIRECTION(1);
+ 	} else
+-		sbp2_prep_command_orb_sg(orb, hi, cmd, scsi_use_sg, sgpnt,
++		sbp2_prep_command_orb_sg(orb, hi, cmd, scsi_use_sg, sg,
+ 					 orb_direction, dma_dir);
+ 
+ 	sbp2util_cpu_to_be32_buffer(orb, sizeof(*orb));
+@@ -1650,7 +1655,6 @@ static int sbp2_send_command(struct sbp2_lu *lu, struct scsi_cmnd *SCpnt,
+ 			     void (*done)(struct scsi_cmnd *))
+ {
+ 	unchar *scsi_cmd = (unchar *)SCpnt->cmnd;
+-	unsigned int request_bufflen = scsi_bufflen(SCpnt);
+ 	struct sbp2_command_info *cmd;
+ 
+ 	cmd = sbp2util_allocate_command_orb(lu, SCpnt, done);
+@@ -1658,7 +1662,7 @@ static int sbp2_send_command(struct sbp2_lu *lu, struct scsi_cmnd *SCpnt,
+ 		return -EIO;
+ 
+ 	sbp2_create_command_orb(lu, cmd, scsi_cmd, scsi_sg_count(SCpnt),
+-				request_bufflen, scsi_sglist(SCpnt),
++				scsi_bufflen(SCpnt), scsi_sglist(SCpnt),
+ 				SCpnt->sc_data_direction);
+ 	sbp2_link_orb_command(lu, cmd);
+ 
+@@ -1981,6 +1985,8 @@ static int sbp2scsi_slave_configure(struct scsi_device *sdev)
+ 		sdev->skip_ms_page_8 = 1;
+ 	if (lu->workarounds & SBP2_WORKAROUND_FIX_CAPACITY)
+ 		sdev->fix_capacity = 1;
++	if (lu->workarounds & SBP2_WORKAROUND_128K_MAX_TRANS)
++		blk_queue_max_sectors(sdev->request_queue, 128 * 1024 / 512);
+ 	return 0;
+ }
+ 
+@@ -2087,9 +2093,6 @@ static int sbp2_module_init(void)
+ 		sbp2_shost_template.cmd_per_lun = 1;
+ 	}
+ 
+-	if (sbp2_default_workarounds & SBP2_WORKAROUND_128K_MAX_TRANS &&
+-	    (sbp2_max_sectors * 512) > (128 * 1024))
+-		sbp2_max_sectors = 128 * 1024 / 512;
+ 	sbp2_shost_template.max_sectors = sbp2_max_sectors;
+ 
+ 	hpsb_register_highlevel(&sbp2_highlevel);
+diff --git a/drivers/ieee1394/sbp2.h b/drivers/ieee1394/sbp2.h
+index 333a4bb..d2ecb0d 100644
+--- a/drivers/ieee1394/sbp2.h
++++ b/drivers/ieee1394/sbp2.h
+@@ -222,7 +222,6 @@ struct sbp2_status_block {
+  */
+ 
+ #define SBP2_MAX_SG_ELEMENT_LENGTH		0xf000
+-#define SBP2_MAX_SECTORS			255
+ /* There is no real limitation of the queue depth (i.e. length of the linked
+  * list of command ORBs) at the target. The chosen depth is merely an
+  * implementation detail of the sbp2 driver. */

Modified: dists/trunk/linux-2.6/debian/patches/series/1~experimental.1
==============================================================================
--- dists/trunk/linux-2.6/debian/patches/series/1~experimental.1	(original)
+++ dists/trunk/linux-2.6/debian/patches/series/1~experimental.1	Mon Dec 24 00:39:41 2007
@@ -36,3 +36,4 @@
 + bugfix/arm/disable-chelsio_t3.patch
 + bugfix/arm/disable-video_bt848.patch
 + bugfix/arm/disable-scsi_acard.patch
++ bugfix/all/git-ieee1394.patch



More information about the Kernel-svn-changes mailing list