UBIFS: fix bulk-read handling uptodate pages

Bulk-read skips uptodate pages but this was putting its
array index out and causing it to treat subsequent pages
as holes.

Signed-off-by: Adrian Hunter <ext-adrian.hunter@nokia.com>
This commit is contained in:
Adrian Hunter 2008-09-12 10:34:51 +03:00 committed by Artem Bityutskiy
parent 46773be497
commit 5c0013c16b

View file

@ -595,7 +595,7 @@ static int ubifs_write_end(struct file *file, struct address_space *mapping,
static int populate_page(struct ubifs_info *c, struct page *page, static int populate_page(struct ubifs_info *c, struct page *page,
struct bu_info *bu, int *n) struct bu_info *bu, int *n)
{ {
int i = 0, nn = *n, offs = bu->zbranch[0].offs, hole = 1, read = 0; int i = 0, nn = *n, offs = bu->zbranch[0].offs, hole = 0, read = 0;
struct inode *inode = page->mapping->host; struct inode *inode = page->mapping->host;
loff_t i_size = i_size_read(inode); loff_t i_size = i_size_read(inode);
unsigned int page_block; unsigned int page_block;
@ -609,6 +609,7 @@ static int populate_page(struct ubifs_info *c, struct page *page,
end_index = (i_size - 1) >> PAGE_CACHE_SHIFT; end_index = (i_size - 1) >> PAGE_CACHE_SHIFT;
if (!i_size || page->index > end_index) { if (!i_size || page->index > end_index) {
hole = 1;
memset(addr, 0, PAGE_CACHE_SIZE); memset(addr, 0, PAGE_CACHE_SIZE);
goto out_hole; goto out_hole;
} }
@ -617,10 +618,10 @@ static int populate_page(struct ubifs_info *c, struct page *page,
while (1) { while (1) {
int err, len, out_len, dlen; int err, len, out_len, dlen;
if (nn >= bu->cnt || if (nn >= bu->cnt) {
key_block(c, &bu->zbranch[nn].key) != page_block) hole = 1;
memset(addr, 0, UBIFS_BLOCK_SIZE); memset(addr, 0, UBIFS_BLOCK_SIZE);
else { } else if (key_block(c, &bu->zbranch[nn].key) == page_block) {
struct ubifs_data_node *dn; struct ubifs_data_node *dn;
dn = bu->buf + (bu->zbranch[nn].offs - offs); dn = bu->buf + (bu->zbranch[nn].offs - offs);
@ -643,8 +644,13 @@ static int populate_page(struct ubifs_info *c, struct page *page,
memset(addr + len, 0, UBIFS_BLOCK_SIZE - len); memset(addr + len, 0, UBIFS_BLOCK_SIZE - len);
nn += 1; nn += 1;
hole = 0;
read = (i << UBIFS_BLOCK_SHIFT) + len; read = (i << UBIFS_BLOCK_SHIFT) + len;
} else if (key_block(c, &bu->zbranch[nn].key) < page_block) {
nn += 1;
continue;
} else {
hole = 1;
memset(addr, 0, UBIFS_BLOCK_SIZE);
} }
if (++i >= UBIFS_BLOCKS_PER_PAGE) if (++i >= UBIFS_BLOCKS_PER_PAGE)
break; break;