[kernel/f17] Apply patches from Jeff Moyer to fix direct-io oops (rhbz 812129)
Josh Boyer
jwboyer at fedoraproject.org
Mon Nov 19 19:17:47 UTC 2012
commit 9ef55ad17118527a9297444f0551f60945e3a2dd
Author: Josh Boyer <jwboyer at redhat.com>
Date: Mon Nov 19 11:04:45 2012 -0500
Apply patches from Jeff Moyer to fix direct-io oops (rhbz 812129)
block-fix-a-crash-when-block-device-is.patch | 214 +++++++++++++++
...-turn-a-rw-semaphore-into-a-percpu-rw-sem.patch | 290 ++++++++++++++++++++
...ck-splice_read-and-splice_write-functions.patch | 74 +++++
kernel.spec | 15 +-
4 files changed, 592 insertions(+), 1 deletions(-)
---
diff --git a/block-fix-a-crash-when-block-device-is.patch b/block-fix-a-crash-when-block-device-is.patch
new file mode 100644
index 0000000..af99283
--- /dev/null
+++ b/block-fix-a-crash-when-block-device-is.patch
@@ -0,0 +1,214 @@
+Fix a crash when block device is read and block size is changed at the same time
+
+commit b87570f5d349661814b262dd5fc40787700f80d6
+Author: Mikulas Patocka <mpatocka at redhat.com>
+Date: Wed Sep 26 07:46:40 2012 +0200
+
+ Fix a crash when block device is read and block size is changed at the same time
+
+ The kernel may crash when block size is changed and I/O is issued
+ simultaneously.
+
+ Because some subsystems (udev or lvm) may read any block device anytime,
+ the bug actually puts any code that changes a block device size in
+ jeopardy.
+
+ The crash can be reproduced if you place "msleep(1000)" to
+ blkdev_get_blocks just before "bh->b_size = max_blocks <<
+ inode->i_blkbits;".
+ Then, run "dd if=/dev/ram0 of=/dev/null bs=4k count=1 iflag=direct"
+ While it is waiting in msleep, run "blockdev --setbsz 2048 /dev/ram0"
+ You get a BUG.
+
+ The direct and non-direct I/O is written with the assumption that block
+ size does not change. It doesn't seem practical to fix these crashes
+ one-by-one there may be many crash possibilities when block size changes
+ at a certain place and it is impossible to find them all and verify the
+ code.
+
+ This patch introduces a new rw-lock bd_block_size_semaphore. The lock is
+ taken for read during I/O. It is taken for write when changing block
+ size. Consequently, block size can't be changed while I/O is being
+ submitted.
+
+ For asynchronous I/O, the patch only prevents block size change while
+ the I/O is being submitted. The block size can change when the I/O is in
+ progress or when the I/O is being finished. This is acceptable because
+ there are no accesses to block size when asynchronous I/O is being
+ finished.
+
+ The patch prevents block size changing while the device is mapped with
+ mmap.
+
+ Signed-off-by: Mikulas Patocka <mpatocka at redhat.com>
+ Signed-off-by: Jens Axboe <axboe at kernel.dk>
+
+Index: linux-3.6.x86_64/drivers/char/raw.c
+===================================================================
+--- linux-3.6.x86_64.orig/drivers/char/raw.c 2012-11-16 17:12:35.127010280 -0500
++++ linux-3.6.x86_64/drivers/char/raw.c 2012-11-16 17:12:37.381002516 -0500
+@@ -285,7 +285,7 @@
+
+ static const struct file_operations raw_fops = {
+ .read = do_sync_read,
+- .aio_read = generic_file_aio_read,
++ .aio_read = blkdev_aio_read,
+ .write = do_sync_write,
+ .aio_write = blkdev_aio_write,
+ .fsync = blkdev_fsync,
+Index: linux-3.6.x86_64/fs/block_dev.c
+===================================================================
+--- linux-3.6.x86_64.orig/fs/block_dev.c 2012-11-16 17:12:35.127010280 -0500
++++ linux-3.6.x86_64/fs/block_dev.c 2012-11-16 17:12:37.381002516 -0500
+@@ -116,6 +116,8 @@
+
+ int set_blocksize(struct block_device *bdev, int size)
+ {
++ struct address_space *mapping;
++
+ /* Size must be a power of two, and between 512 and PAGE_SIZE */
+ if (size > PAGE_SIZE || size < 512 || !is_power_of_2(size))
+ return -EINVAL;
+@@ -124,6 +126,20 @@
+ if (size < bdev_logical_block_size(bdev))
+ return -EINVAL;
+
++ /* Prevent starting I/O or mapping the device */
++ down_write(&bdev->bd_block_size_semaphore);
++
++ /* Check that the block device is not memory mapped */
++ mapping = bdev->bd_inode->i_mapping;
++ mutex_lock(&mapping->i_mmap_mutex);
++ if (!prio_tree_empty(&mapping->i_mmap) ||
++ !list_empty(&mapping->i_mmap_nonlinear)) {
++ mutex_unlock(&mapping->i_mmap_mutex);
++ up_write(&bdev->bd_block_size_semaphore);
++ return -EBUSY;
++ }
++ mutex_unlock(&mapping->i_mmap_mutex);
++
+ /* Don't change the size if it is same as current */
+ if (bdev->bd_block_size != size) {
+ sync_blockdev(bdev);
+@@ -131,6 +147,9 @@
+ bdev->bd_inode->i_blkbits = blksize_bits(size);
+ kill_bdev(bdev);
+ }
++
++ up_write(&bdev->bd_block_size_semaphore);
++
+ return 0;
+ }
+
+@@ -472,6 +491,7 @@
+ inode_init_once(&ei->vfs_inode);
+ /* Initialize mutex for freeze. */
+ mutex_init(&bdev->bd_fsfreeze_mutex);
++ init_rwsem(&bdev->bd_block_size_semaphore);
+ }
+
+ static inline void __bd_forget(struct inode *inode)
+@@ -1567,6 +1587,22 @@
+ return blkdev_ioctl(bdev, mode, cmd, arg);
+ }
+
++ssize_t blkdev_aio_read(struct kiocb *iocb, const struct iovec *iov,
++ unsigned long nr_segs, loff_t pos)
++{
++ ssize_t ret;
++ struct block_device *bdev = I_BDEV(iocb->ki_filp->f_mapping->host);
++
++ down_read(&bdev->bd_block_size_semaphore);
++
++ ret = generic_file_aio_read(iocb, iov, nr_segs, pos);
++
++ up_read(&bdev->bd_block_size_semaphore);
++
++ return ret;
++}
++EXPORT_SYMBOL_GPL(blkdev_aio_read);
++
+ /*
+ * Write data to the block device. Only intended for the block device itself
+ * and the raw driver which basically is a fake block device.
+@@ -1578,12 +1614,16 @@
+ unsigned long nr_segs, loff_t pos)
+ {
+ struct file *file = iocb->ki_filp;
++ struct block_device *bdev = I_BDEV(file->f_mapping->host);
+ struct blk_plug plug;
+ ssize_t ret;
+
+ BUG_ON(iocb->ki_pos != pos);
+
+ blk_start_plug(&plug);
++
++ down_read(&bdev->bd_block_size_semaphore);
++
+ ret = __generic_file_aio_write(iocb, iov, nr_segs, &iocb->ki_pos);
+ if (ret > 0 || ret == -EIOCBQUEUED) {
+ ssize_t err;
+@@ -1592,11 +1632,29 @@
+ if (err < 0 && ret > 0)
+ ret = err;
+ }
++
++ up_read(&bdev->bd_block_size_semaphore);
++
+ blk_finish_plug(&plug);
++
+ return ret;
+ }
+ EXPORT_SYMBOL_GPL(blkdev_aio_write);
+
++int blkdev_mmap(struct file *file, struct vm_area_struct *vma)
++{
++ int ret;
++ struct block_device *bdev = I_BDEV(file->f_mapping->host);
++
++ down_read(&bdev->bd_block_size_semaphore);
++
++ ret = generic_file_mmap(file, vma);
++
++ up_read(&bdev->bd_block_size_semaphore);
++
++ return ret;
++}
++
+ /*
+ * Try to release a page associated with block device when the system
+ * is under memory pressure.
+@@ -1627,9 +1685,9 @@
+ .llseek = block_llseek,
+ .read = do_sync_read,
+ .write = do_sync_write,
+- .aio_read = generic_file_aio_read,
++ .aio_read = blkdev_aio_read,
+ .aio_write = blkdev_aio_write,
+- .mmap = generic_file_mmap,
++ .mmap = blkdev_mmap,
+ .fsync = blkdev_fsync,
+ .unlocked_ioctl = block_ioctl,
+ #ifdef CONFIG_COMPAT
+Index: linux-3.6.x86_64/include/linux/fs.h
+===================================================================
+--- linux-3.6.x86_64.orig/include/linux/fs.h 2012-11-16 17:12:35.127010280 -0500
++++ linux-3.6.x86_64/include/linux/fs.h 2012-11-16 17:12:37.424002387 -0500
+@@ -724,6 +724,8 @@
+ int bd_fsfreeze_count;
+ /* Mutex for freeze */
+ struct mutex bd_fsfreeze_mutex;
++ /* A semaphore that prevents I/O while block size is being changed */
++ struct rw_semaphore bd_block_size_semaphore;
+ };
+
+ /*
+@@ -2564,6 +2566,8 @@
+ unsigned long *nr_segs, size_t *count, int access_flags);
+
+ /* fs/block_dev.c */
++extern ssize_t blkdev_aio_read(struct kiocb *iocb, const struct iovec *iov,
++ unsigned long nr_segs, loff_t pos);
+ extern ssize_t blkdev_aio_write(struct kiocb *iocb, const struct iovec *iov,
+ unsigned long nr_segs, loff_t pos);
+ extern int blkdev_fsync(struct file *filp, loff_t start, loff_t end,
diff --git a/blockdev-turn-a-rw-semaphore-into-a-percpu-rw-sem.patch b/blockdev-turn-a-rw-semaphore-into-a-percpu-rw-sem.patch
new file mode 100644
index 0000000..82caa6b
--- /dev/null
+++ b/blockdev-turn-a-rw-semaphore-into-a-percpu-rw-sem.patch
@@ -0,0 +1,290 @@
+blockdev: turn a rw semaphore into a percpu rw semaphore
+
+commit 62ac665ff9fc07497ca524bd20d6a96893d11071
+Author: Mikulas Patocka <mpatocka at redhat.com>
+Date: Wed Sep 26 07:46:43 2012 +0200
+
+ blockdev: turn a rw semaphore into a percpu rw semaphore
+
+ This avoids cache line bouncing when many processes lock the semaphore
+ for read.
+
+ New percpu lock implementation
+
+ The lock consists of an array of percpu unsigned integers, a boolean
+ variable and a mutex.
+
+ When we take the lock for read, we enter rcu read section, check for a
+ "locked" variable. If it is false, we increase a percpu counter on the
+ current cpu and exit the rcu section. If "locked" is true, we exit the
+ rcu section, take the mutex and drop it (this waits until a writer
+ finished) and retry.
+
+ Unlocking for read just decreases percpu variable. Note that we can
+ unlock on a difference cpu than where we locked, in this case the
+ counter underflows. The sum of all percpu counters represents the number
+ of processes that hold the lock for read.
+
+ When we need to lock for write, we take the mutex, set "locked" variable
+ to true and synchronize rcu. Since RCU has been synchronized, no
+ processes can create new read locks. We wait until the sum of percpu
+ counters is zero - when it is, there are no readers in the critical
+ section.
+
+ Signed-off-by: Mikulas Patocka <mpatocka at redhat.com>
+ Signed-off-by: Jens Axboe <axboe at kernel.dk>
+
+Index: linux-3.6.x86_64/Documentation/percpu-rw-semaphore.txt
+===================================================================
+--- /dev/null 1970-01-01 00:00:00.000000000 +0000
++++ linux-3.6.x86_64/Documentation/percpu-rw-semaphore.txt 2012-11-16 17:12:57.351936583 -0500
+@@ -0,0 +1,27 @@
++Percpu rw semaphores
++--------------------
++
++Percpu rw semaphores is a new read-write semaphore design that is
++optimized for locking for reading.
++
++The problem with traditional read-write semaphores is that when multiple
++cores take the lock for reading, the cache line containing the semaphore
++is bouncing between L1 caches of the cores, causing performance
++degradation.
++
++Locking for reading it very fast, it uses RCU and it avoids any atomic
++instruction in the lock and unlock path. On the other hand, locking for
++writing is very expensive, it calls synchronize_rcu() that can take
++hundreds of microseconds.
++
++The lock is declared with "struct percpu_rw_semaphore" type.
++The lock is initialized percpu_init_rwsem, it returns 0 on success and
++-ENOMEM on allocation failure.
++The lock must be freed with percpu_free_rwsem to avoid memory leak.
++
++The lock is locked for read with percpu_down_read, percpu_up_read and
++for write with percpu_down_write, percpu_up_write.
++
++The idea of using RCU for optimized rw-lock was introduced by
++Eric Dumazet <eric.dumazet at gmail.com>.
++The code was written by Mikulas Patocka <mpatocka at redhat.com>
+Index: linux-3.6.x86_64/fs/block_dev.c
+===================================================================
+--- linux-3.6.x86_64.orig/fs/block_dev.c 2012-11-16 17:12:37.381002516 -0500
++++ linux-3.6.x86_64/fs/block_dev.c 2012-11-16 17:27:41.217005828 -0500
+@@ -127,7 +127,7 @@
+ return -EINVAL;
+
+ /* Prevent starting I/O or mapping the device */
+- down_write(&bdev->bd_block_size_semaphore);
++ percpu_down_write(&bdev->bd_block_size_semaphore);
+
+ /* Check that the block device is not memory mapped */
+ mapping = bdev->bd_inode->i_mapping;
+@@ -135,7 +135,7 @@
+ if (!prio_tree_empty(&mapping->i_mmap) ||
+ !list_empty(&mapping->i_mmap_nonlinear)) {
+ mutex_unlock(&mapping->i_mmap_mutex);
+- up_write(&bdev->bd_block_size_semaphore);
++ percpu_up_write(&bdev->bd_block_size_semaphore);
+ return -EBUSY;
+ }
+ mutex_unlock(&mapping->i_mmap_mutex);
+@@ -148,7 +148,7 @@
+ kill_bdev(bdev);
+ }
+
+- up_write(&bdev->bd_block_size_semaphore);
++ percpu_up_write(&bdev->bd_block_size_semaphore);
+
+ return 0;
+ }
+@@ -460,6 +460,12 @@
+ struct bdev_inode *ei = kmem_cache_alloc(bdev_cachep, GFP_KERNEL);
+ if (!ei)
+ return NULL;
++
++ if (unlikely(percpu_init_rwsem(&ei->bdev.bd_block_size_semaphore))) {
++ kmem_cache_free(bdev_cachep, ei);
++ return NULL;
++ }
++
+ return &ei->vfs_inode;
+ }
+
+@@ -468,6 +474,8 @@
+ struct inode *inode = container_of(head, struct inode, i_rcu);
+ struct bdev_inode *bdi = BDEV_I(inode);
+
++ percpu_free_rwsem(&bdi->bdev.bd_block_size_semaphore);
++
+ kmem_cache_free(bdev_cachep, bdi);
+ }
+
+@@ -491,7 +499,6 @@
+ inode_init_once(&ei->vfs_inode);
+ /* Initialize mutex for freeze. */
+ mutex_init(&bdev->bd_fsfreeze_mutex);
+- init_rwsem(&bdev->bd_block_size_semaphore);
+ }
+
+ static inline void __bd_forget(struct inode *inode)
+@@ -1593,11 +1600,11 @@
+ ssize_t ret;
+ struct block_device *bdev = I_BDEV(iocb->ki_filp->f_mapping->host);
+
+- down_read(&bdev->bd_block_size_semaphore);
++ percpu_down_read(&bdev->bd_block_size_semaphore);
+
+ ret = generic_file_aio_read(iocb, iov, nr_segs, pos);
+
+- up_read(&bdev->bd_block_size_semaphore);
++ percpu_up_read(&bdev->bd_block_size_semaphore);
+
+ return ret;
+ }
+@@ -1622,7 +1629,7 @@
+
+ blk_start_plug(&plug);
+
+- down_read(&bdev->bd_block_size_semaphore);
++ percpu_down_read(&bdev->bd_block_size_semaphore);
+
+ ret = __generic_file_aio_write(iocb, iov, nr_segs, &iocb->ki_pos);
+ if (ret > 0 || ret == -EIOCBQUEUED) {
+@@ -1633,7 +1640,7 @@
+ ret = err;
+ }
+
+- up_read(&bdev->bd_block_size_semaphore);
++ percpu_up_read(&bdev->bd_block_size_semaphore);
+
+ blk_finish_plug(&plug);
+
+@@ -1646,11 +1653,11 @@
+ int ret;
+ struct block_device *bdev = I_BDEV(file->f_mapping->host);
+
+- down_read(&bdev->bd_block_size_semaphore);
++ percpu_down_read(&bdev->bd_block_size_semaphore);
+
+ ret = generic_file_mmap(file, vma);
+
+- up_read(&bdev->bd_block_size_semaphore);
++ percpu_up_read(&bdev->bd_block_size_semaphore);
+
+ return ret;
+ }
+Index: linux-3.6.x86_64/include/linux/fs.h
+===================================================================
+--- linux-3.6.x86_64.orig/include/linux/fs.h 2012-11-16 17:12:37.424002387 -0500
++++ linux-3.6.x86_64/include/linux/fs.h 2012-11-16 17:28:12.578901349 -0500
+@@ -415,6 +415,7 @@
+ #include <linux/migrate_mode.h>
+ #include <linux/uidgid.h>
+ #include <linux/lockdep.h>
++#include <linux/percpu-rwsem.h>
+
+ #include <asm/byteorder.h>
+
+@@ -725,7 +726,7 @@
+ /* Mutex for freeze */
+ struct mutex bd_fsfreeze_mutex;
+ /* A semaphore that prevents I/O while block size is being changed */
+- struct rw_semaphore bd_block_size_semaphore;
++ struct percpu_rw_semaphore bd_block_size_semaphore;
+ };
+
+ /*
+Index: linux-3.6.x86_64/include/linux/percpu-rwsem.h
+===================================================================
+--- /dev/null 1970-01-01 00:00:00.000000000 +0000
++++ linux-3.6.x86_64/include/linux/percpu-rwsem.h 2012-11-16 17:12:57.354936574 -0500
+@@ -0,0 +1,89 @@
++#ifndef _LINUX_PERCPU_RWSEM_H
++#define _LINUX_PERCPU_RWSEM_H
++
++#include <linux/mutex.h>
++#include <linux/percpu.h>
++#include <linux/rcupdate.h>
++#include <linux/delay.h>
++
++struct percpu_rw_semaphore {
++ unsigned __percpu *counters;
++ bool locked;
++ struct mutex mtx;
++};
++
++static inline void percpu_down_read(struct percpu_rw_semaphore *p)
++{
++ rcu_read_lock();
++ if (unlikely(p->locked)) {
++ rcu_read_unlock();
++ mutex_lock(&p->mtx);
++ this_cpu_inc(*p->counters);
++ mutex_unlock(&p->mtx);
++ return;
++ }
++ this_cpu_inc(*p->counters);
++ rcu_read_unlock();
++}
++
++static inline void percpu_up_read(struct percpu_rw_semaphore *p)
++{
++ /*
++ * On X86, write operation in this_cpu_dec serves as a memory unlock
++ * barrier (i.e. memory accesses may be moved before the write, but
++ * no memory accesses are moved past the write).
++ * On other architectures this may not be the case, so we need smp_mb()
++ * there.
++ */
++#if defined(CONFIG_X86) && (!defined(CONFIG_X86_PPRO_FENCE) && !defined(CONFIG_X86_OOSTORE))
++ barrier();
++#else
++ smp_mb();
++#endif
++ this_cpu_dec(*p->counters);
++}
++
++static inline unsigned __percpu_count(unsigned __percpu *counters)
++{
++ unsigned total = 0;
++ int cpu;
++
++ for_each_possible_cpu(cpu)
++ total += ACCESS_ONCE(*per_cpu_ptr(counters, cpu));
++
++ return total;
++}
++
++static inline void percpu_down_write(struct percpu_rw_semaphore *p)
++{
++ mutex_lock(&p->mtx);
++ p->locked = true;
++ synchronize_rcu();
++ while (__percpu_count(p->counters))
++ msleep(1);
++ smp_rmb(); /* paired with smp_mb() in percpu_sem_up_read() */
++}
++
++static inline void percpu_up_write(struct percpu_rw_semaphore *p)
++{
++ p->locked = false;
++ mutex_unlock(&p->mtx);
++}
++
++static inline int percpu_init_rwsem(struct percpu_rw_semaphore *p)
++{
++ p->counters = alloc_percpu(unsigned);
++ if (unlikely(!p->counters))
++ return -ENOMEM;
++ p->locked = false;
++ mutex_init(&p->mtx);
++ return 0;
++}
++
++static inline void percpu_free_rwsem(struct percpu_rw_semaphore *p)
++{
++ free_percpu(p->counters);
++ p->counters = NULL; /* catch use after free bugs */
++}
++
++#endif
diff --git a/fs-lock-splice_read-and-splice_write-functions.patch b/fs-lock-splice_read-and-splice_write-functions.patch
new file mode 100644
index 0000000..938cbd9
--- /dev/null
+++ b/fs-lock-splice_read-and-splice_write-functions.patch
@@ -0,0 +1,74 @@
+Lock splice_read and splice_write functions
+
+commit 1a25b1c4ce189e3926f2981f3302352a930086db
+Author: Mikulas Patocka <mpatocka at redhat.com>
+Date: Mon Oct 15 17:20:17 2012 -0400
+
+ Lock splice_read and splice_write functions
+
+ Functions generic_file_splice_read and generic_file_splice_write access
+ the pagecache directly. For block devices these functions must be locked
+ so that block size is not changed while they are in progress.
+
+ This patch is an additional fix for commit b87570f5d349 ("Fix a crash
+ when block device is read and block size is changed at the same time")
+ that locked aio_read, aio_write and mmap against block size change.
+
+ Signed-off-by: Mikulas Patocka <mpatocka at redhat.com>
+ Signed-off-by: Linus Torvalds <torvalds at linux-foundation.org>
+
+Index: linux-3.6.x86_64/fs/block_dev.c
+===================================================================
+--- linux-3.6.x86_64.orig/fs/block_dev.c 2012-11-16 17:12:57.352936580 -0500
++++ linux-3.6.x86_64/fs/block_dev.c 2012-11-16 17:13:11.908887989 -0500
+@@ -1662,6 +1662,39 @@
+ return ret;
+ }
+
++static ssize_t blkdev_splice_read(struct file *file, loff_t *ppos,
++ struct pipe_inode_info *pipe, size_t len,
++ unsigned int flags)
++{
++ ssize_t ret;
++ struct block_device *bdev = I_BDEV(file->f_mapping->host);
++
++ percpu_down_read(&bdev->bd_block_size_semaphore);
++
++ ret = generic_file_splice_read(file, ppos, pipe, len, flags);
++
++ percpu_up_read(&bdev->bd_block_size_semaphore);
++
++ return ret;
++}
++
++static ssize_t blkdev_splice_write(struct pipe_inode_info *pipe,
++ struct file *file, loff_t *ppos, size_t len,
++ unsigned int flags)
++{
++ ssize_t ret;
++ struct block_device *bdev = I_BDEV(file->f_mapping->host);
++
++ percpu_down_read(&bdev->bd_block_size_semaphore);
++
++ ret = generic_file_splice_write(pipe, file, ppos, len, flags);
++
++ percpu_up_read(&bdev->bd_block_size_semaphore);
++
++ return ret;
++}
++
++
+ /*
+ * Try to release a page associated with block device when the system
+ * is under memory pressure.
+@@ -1700,8 +1733,8 @@
+ #ifdef CONFIG_COMPAT
+ .compat_ioctl = compat_blkdev_ioctl,
+ #endif
+- .splice_read = generic_file_splice_read,
+- .splice_write = generic_file_splice_write,
++ .splice_read = blkdev_splice_read,
++ .splice_write = blkdev_splice_write,
+ };
+
+ int ioctl_by_bdev(struct block_device *bdev, unsigned cmd, unsigned long arg)
diff --git a/kernel.spec b/kernel.spec
index 8e179c2..be512c7 100644
--- a/kernel.spec
+++ b/kernel.spec
@@ -54,7 +54,7 @@ Summary: The Linux kernel
# For non-released -rc kernels, this will be appended after the rcX and
# gitX tags, so a 3 here would become part of release "0.rcX.gitX.3"
#
-%global baserelease 1
+%global baserelease 2
%global fedora_build %{baserelease}
# base_sublevel is the kernel version we're starting with and patching
@@ -779,6 +779,11 @@ Patch22114: iwlwifi-remove-queue-empty-warn-3.6.patch
#rhbz 870562
Patch22115: keyspan.patch
+#rhbz 812129
+Patch22120: block-fix-a-crash-when-block-device-is.patch
+Patch22121: blockdev-turn-a-rw-semaphore-into-a-percpu-rw-sem.patch
+Patch22122: fs-lock-splice_read-and-splice_write-functions.patch
+
# END OF PATCH DEFINITIONS
%endif
@@ -1507,6 +1512,11 @@ ApplyPatch iwlwifi-remove-queue-empty-warn-3.6.patch
#rhbz 870562
ApplyPatch keyspan.patch
+#rhbz 812129
+ApplyPatch block-fix-a-crash-when-block-device-is.patch
+ApplyPatch blockdev-turn-a-rw-semaphore-into-a-percpu-rw-sem.patch
+ApplyPatch fs-lock-splice_read-and-splice_write-functions.patch
+
# END OF PATCH APPLICATIONS
%endif
@@ -2370,6 +2380,9 @@ fi
# '-' | |
# '-'
%changelog
+* Mon Nov 19 2012 Josh Boyer <jwboyer at redhat.com>
+- Apply patches from Jeff Moyer to fix direct-io oops (rhbz 812129)
+
* Sat Nov 17 2012 Justin M. Forbes <jforbes at linuxtx.org> - 3.6.7-1
- linux 3.6.7
More information about the scm-commits
mailing list