 1f5abe7e7d
			
		
	
	
	1f5abe7e7d
	
	
	
		
			
			Pekka Enberg advised me: > It would be nice if BUG(), BUG_ON(), and panic() calls would be > converted to proper error handling using WARN_ON() calls. The BUG() > call in nilfs_cpfile_delete_checkpoints(), for example, looks to be > triggerable from user-space via the ioctl() system call. This will follow the comment and keep them to a minimum. Acked-by: Pekka Enberg <penberg@cs.helsinki.fi> Signed-off-by: Ryusuke Konishi <konishi.ryusuke@lab.ntt.co.jp> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
		
			
				
	
	
		
			430 lines
		
	
	
	
		
			12 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			430 lines
		
	
	
	
		
			12 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * dat.c - NILFS disk address translation.
 | |
|  *
 | |
|  * Copyright (C) 2006-2008 Nippon Telegraph and Telephone Corporation.
 | |
|  *
 | |
|  * This program is free software; you can redistribute it and/or modify
 | |
|  * it under the terms of the GNU General Public License as published by
 | |
|  * the Free Software Foundation; either version 2 of the License, or
 | |
|  * (at your option) any later version.
 | |
|  *
 | |
|  * This program is distributed in the hope that it will be useful,
 | |
|  * but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
|  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 | |
|  * GNU General Public License for more details.
 | |
|  *
 | |
|  * You should have received a copy of the GNU General Public License
 | |
|  * along with this program; if not, write to the Free Software
 | |
|  * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA  02110-1301  USA
 | |
|  *
 | |
|  * Written by Koji Sato <koji@osrg.net>.
 | |
|  */
 | |
| 
 | |
| #include <linux/types.h>
 | |
| #include <linux/buffer_head.h>
 | |
| #include <linux/string.h>
 | |
| #include <linux/errno.h>
 | |
| #include "nilfs.h"
 | |
| #include "mdt.h"
 | |
| #include "alloc.h"
 | |
| #include "dat.h"
 | |
| 
 | |
| 
 | |
| #define NILFS_CNO_MIN	((__u64)1)
 | |
| #define NILFS_CNO_MAX	(~(__u64)0)
 | |
| 
 | |
| static int nilfs_dat_prepare_entry(struct inode *dat,
 | |
| 				   struct nilfs_palloc_req *req, int create)
 | |
| {
 | |
| 	return nilfs_palloc_get_entry_block(dat, req->pr_entry_nr,
 | |
| 					    create, &req->pr_entry_bh);
 | |
| }
 | |
| 
 | |
| static void nilfs_dat_commit_entry(struct inode *dat,
 | |
| 				   struct nilfs_palloc_req *req)
 | |
| {
 | |
| 	nilfs_mdt_mark_buffer_dirty(req->pr_entry_bh);
 | |
| 	nilfs_mdt_mark_dirty(dat);
 | |
| 	brelse(req->pr_entry_bh);
 | |
| }
 | |
| 
 | |
| static void nilfs_dat_abort_entry(struct inode *dat,
 | |
| 				  struct nilfs_palloc_req *req)
 | |
| {
 | |
| 	brelse(req->pr_entry_bh);
 | |
| }
 | |
| 
 | |
| int nilfs_dat_prepare_alloc(struct inode *dat, struct nilfs_palloc_req *req)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	ret = nilfs_palloc_prepare_alloc_entry(dat, req);
 | |
| 	if (ret < 0)
 | |
| 		return ret;
 | |
| 
 | |
| 	ret = nilfs_dat_prepare_entry(dat, req, 1);
 | |
| 	if (ret < 0)
 | |
| 		nilfs_palloc_abort_alloc_entry(dat, req);
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| void nilfs_dat_commit_alloc(struct inode *dat, struct nilfs_palloc_req *req)
 | |
| {
 | |
| 	struct nilfs_dat_entry *entry;
 | |
| 	void *kaddr;
 | |
| 
 | |
| 	kaddr = kmap_atomic(req->pr_entry_bh->b_page, KM_USER0);
 | |
| 	entry = nilfs_palloc_block_get_entry(dat, req->pr_entry_nr,
 | |
| 					     req->pr_entry_bh, kaddr);
 | |
| 	entry->de_start = cpu_to_le64(NILFS_CNO_MIN);
 | |
| 	entry->de_end = cpu_to_le64(NILFS_CNO_MAX);
 | |
| 	entry->de_blocknr = cpu_to_le64(0);
 | |
| 	kunmap_atomic(kaddr, KM_USER0);
 | |
| 
 | |
| 	nilfs_palloc_commit_alloc_entry(dat, req);
 | |
| 	nilfs_dat_commit_entry(dat, req);
 | |
| }
 | |
| 
 | |
| void nilfs_dat_abort_alloc(struct inode *dat, struct nilfs_palloc_req *req)
 | |
| {
 | |
| 	nilfs_dat_abort_entry(dat, req);
 | |
| 	nilfs_palloc_abort_alloc_entry(dat, req);
 | |
| }
 | |
| 
 | |
| int nilfs_dat_prepare_free(struct inode *dat, struct nilfs_palloc_req *req)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	ret = nilfs_palloc_prepare_free_entry(dat, req);
 | |
| 	if (ret < 0)
 | |
| 		return ret;
 | |
| 	ret = nilfs_dat_prepare_entry(dat, req, 0);
 | |
| 	if (ret < 0) {
 | |
| 		nilfs_palloc_abort_free_entry(dat, req);
 | |
| 		return ret;
 | |
| 	}
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| void nilfs_dat_commit_free(struct inode *dat, struct nilfs_palloc_req *req)
 | |
| {
 | |
| 	struct nilfs_dat_entry *entry;
 | |
| 	void *kaddr;
 | |
| 
 | |
| 	kaddr = kmap_atomic(req->pr_entry_bh->b_page, KM_USER0);
 | |
| 	entry = nilfs_palloc_block_get_entry(dat, req->pr_entry_nr,
 | |
| 					     req->pr_entry_bh, kaddr);
 | |
| 	entry->de_start = cpu_to_le64(NILFS_CNO_MIN);
 | |
| 	entry->de_end = cpu_to_le64(NILFS_CNO_MIN);
 | |
| 	entry->de_blocknr = cpu_to_le64(0);
 | |
| 	kunmap_atomic(kaddr, KM_USER0);
 | |
| 
 | |
| 	nilfs_dat_commit_entry(dat, req);
 | |
| 	nilfs_palloc_commit_free_entry(dat, req);
 | |
| }
 | |
| 
 | |
| void nilfs_dat_abort_free(struct inode *dat, struct nilfs_palloc_req *req)
 | |
| {
 | |
| 	nilfs_dat_abort_entry(dat, req);
 | |
| 	nilfs_palloc_abort_free_entry(dat, req);
 | |
| }
 | |
| 
 | |
| int nilfs_dat_prepare_start(struct inode *dat, struct nilfs_palloc_req *req)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	ret = nilfs_dat_prepare_entry(dat, req, 0);
 | |
| 	WARN_ON(ret == -ENOENT);
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| void nilfs_dat_commit_start(struct inode *dat, struct nilfs_palloc_req *req,
 | |
| 			    sector_t blocknr)
 | |
| {
 | |
| 	struct nilfs_dat_entry *entry;
 | |
| 	void *kaddr;
 | |
| 
 | |
| 	kaddr = kmap_atomic(req->pr_entry_bh->b_page, KM_USER0);
 | |
| 	entry = nilfs_palloc_block_get_entry(dat, req->pr_entry_nr,
 | |
| 					     req->pr_entry_bh, kaddr);
 | |
| 	entry->de_start = cpu_to_le64(nilfs_mdt_cno(dat));
 | |
| 	if (entry->de_blocknr != cpu_to_le64(0) ||
 | |
| 	    entry->de_end != cpu_to_le64(NILFS_CNO_MAX)) {
 | |
| 		printk(KERN_CRIT
 | |
| 		       "%s: vbn = %llu, start = %llu, end = %llu, pbn = %llu\n",
 | |
| 		       __func__, (unsigned long long)req->pr_entry_nr,
 | |
| 		       (unsigned long long)le64_to_cpu(entry->de_start),
 | |
| 		       (unsigned long long)le64_to_cpu(entry->de_end),
 | |
| 		       (unsigned long long)le64_to_cpu(entry->de_blocknr));
 | |
| 	}
 | |
| 	entry->de_blocknr = cpu_to_le64(blocknr);
 | |
| 	kunmap_atomic(kaddr, KM_USER0);
 | |
| 
 | |
| 	nilfs_dat_commit_entry(dat, req);
 | |
| }
 | |
| 
 | |
| void nilfs_dat_abort_start(struct inode *dat, struct nilfs_palloc_req *req)
 | |
| {
 | |
| 	nilfs_dat_abort_entry(dat, req);
 | |
| }
 | |
| 
 | |
| int nilfs_dat_prepare_end(struct inode *dat, struct nilfs_palloc_req *req)
 | |
| {
 | |
| 	struct nilfs_dat_entry *entry;
 | |
| 	__u64 start;
 | |
| 	sector_t blocknr;
 | |
| 	void *kaddr;
 | |
| 	int ret;
 | |
| 
 | |
| 	ret = nilfs_dat_prepare_entry(dat, req, 0);
 | |
| 	if (ret < 0) {
 | |
| 		WARN_ON(ret == -ENOENT);
 | |
| 		return ret;
 | |
| 	}
 | |
| 
 | |
| 	kaddr = kmap_atomic(req->pr_entry_bh->b_page, KM_USER0);
 | |
| 	entry = nilfs_palloc_block_get_entry(dat, req->pr_entry_nr,
 | |
| 					     req->pr_entry_bh, kaddr);
 | |
| 	start = le64_to_cpu(entry->de_start);
 | |
| 	blocknr = le64_to_cpu(entry->de_blocknr);
 | |
| 	kunmap_atomic(kaddr, KM_USER0);
 | |
| 
 | |
| 	if (blocknr == 0) {
 | |
| 		ret = nilfs_palloc_prepare_free_entry(dat, req);
 | |
| 		if (ret < 0) {
 | |
| 			nilfs_dat_abort_entry(dat, req);
 | |
| 			return ret;
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| void nilfs_dat_commit_end(struct inode *dat, struct nilfs_palloc_req *req,
 | |
| 			  int dead)
 | |
| {
 | |
| 	struct nilfs_dat_entry *entry;
 | |
| 	__u64 start, end;
 | |
| 	sector_t blocknr;
 | |
| 	void *kaddr;
 | |
| 
 | |
| 	kaddr = kmap_atomic(req->pr_entry_bh->b_page, KM_USER0);
 | |
| 	entry = nilfs_palloc_block_get_entry(dat, req->pr_entry_nr,
 | |
| 					     req->pr_entry_bh, kaddr);
 | |
| 	end = start = le64_to_cpu(entry->de_start);
 | |
| 	if (!dead) {
 | |
| 		end = nilfs_mdt_cno(dat);
 | |
| 		WARN_ON(start > end);
 | |
| 	}
 | |
| 	entry->de_end = cpu_to_le64(end);
 | |
| 	blocknr = le64_to_cpu(entry->de_blocknr);
 | |
| 	kunmap_atomic(kaddr, KM_USER0);
 | |
| 
 | |
| 	if (blocknr == 0)
 | |
| 		nilfs_dat_commit_free(dat, req);
 | |
| 	else
 | |
| 		nilfs_dat_commit_entry(dat, req);
 | |
| }
 | |
| 
 | |
| void nilfs_dat_abort_end(struct inode *dat, struct nilfs_palloc_req *req)
 | |
| {
 | |
| 	struct nilfs_dat_entry *entry;
 | |
| 	__u64 start;
 | |
| 	sector_t blocknr;
 | |
| 	void *kaddr;
 | |
| 
 | |
| 	kaddr = kmap_atomic(req->pr_entry_bh->b_page, KM_USER0);
 | |
| 	entry = nilfs_palloc_block_get_entry(dat, req->pr_entry_nr,
 | |
| 					     req->pr_entry_bh, kaddr);
 | |
| 	start = le64_to_cpu(entry->de_start);
 | |
| 	blocknr = le64_to_cpu(entry->de_blocknr);
 | |
| 	kunmap_atomic(kaddr, KM_USER0);
 | |
| 
 | |
| 	if (start == nilfs_mdt_cno(dat) && blocknr == 0)
 | |
| 		nilfs_palloc_abort_free_entry(dat, req);
 | |
| 	nilfs_dat_abort_entry(dat, req);
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * nilfs_dat_mark_dirty -
 | |
|  * @dat: DAT file inode
 | |
|  * @vblocknr: virtual block number
 | |
|  *
 | |
|  * Description:
 | |
|  *
 | |
|  * Return Value: On success, 0 is returned. On error, one of the following
 | |
|  * negative error codes is returned.
 | |
|  *
 | |
|  * %-EIO - I/O error.
 | |
|  *
 | |
|  * %-ENOMEM - Insufficient amount of memory available.
 | |
|  */
 | |
| int nilfs_dat_mark_dirty(struct inode *dat, __u64 vblocknr)
 | |
| {
 | |
| 	struct nilfs_palloc_req req;
 | |
| 	int ret;
 | |
| 
 | |
| 	req.pr_entry_nr = vblocknr;
 | |
| 	ret = nilfs_dat_prepare_entry(dat, &req, 0);
 | |
| 	if (ret == 0)
 | |
| 		nilfs_dat_commit_entry(dat, &req);
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * nilfs_dat_freev - free virtual block numbers
 | |
|  * @dat: DAT file inode
 | |
|  * @vblocknrs: array of virtual block numbers
 | |
|  * @nitems: number of virtual block numbers
 | |
|  *
 | |
|  * Description: nilfs_dat_freev() frees the virtual block numbers specified by
 | |
|  * @vblocknrs and @nitems.
 | |
|  *
 | |
|  * Return Value: On success, 0 is returned. On error, one of the following
 | |
|  * nagative error codes is returned.
 | |
|  *
 | |
|  * %-EIO - I/O error.
 | |
|  *
 | |
|  * %-ENOMEM - Insufficient amount of memory available.
 | |
|  *
 | |
|  * %-ENOENT - The virtual block number have not been allocated.
 | |
|  */
 | |
| int nilfs_dat_freev(struct inode *dat, __u64 *vblocknrs, size_t nitems)
 | |
| {
 | |
| 	return nilfs_palloc_freev(dat, vblocknrs, nitems);
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * nilfs_dat_move - change a block number
 | |
|  * @dat: DAT file inode
 | |
|  * @vblocknr: virtual block number
 | |
|  * @blocknr: block number
 | |
|  *
 | |
|  * Description: nilfs_dat_move() changes the block number associated with
 | |
|  * @vblocknr to @blocknr.
 | |
|  *
 | |
|  * Return Value: On success, 0 is returned. On error, one of the following
 | |
|  * negative error codes is returned.
 | |
|  *
 | |
|  * %-EIO - I/O error.
 | |
|  *
 | |
|  * %-ENOMEM - Insufficient amount of memory available.
 | |
|  */
 | |
| int nilfs_dat_move(struct inode *dat, __u64 vblocknr, sector_t blocknr)
 | |
| {
 | |
| 	struct buffer_head *entry_bh;
 | |
| 	struct nilfs_dat_entry *entry;
 | |
| 	void *kaddr;
 | |
| 	int ret;
 | |
| 
 | |
| 	ret = nilfs_palloc_get_entry_block(dat, vblocknr, 0, &entry_bh);
 | |
| 	if (ret < 0)
 | |
| 		return ret;
 | |
| 	kaddr = kmap_atomic(entry_bh->b_page, KM_USER0);
 | |
| 	entry = nilfs_palloc_block_get_entry(dat, vblocknr, entry_bh, kaddr);
 | |
| 	if (unlikely(entry->de_blocknr == cpu_to_le64(0))) {
 | |
| 		printk(KERN_CRIT "%s: vbn = %llu, [%llu, %llu)\n", __func__,
 | |
| 		       (unsigned long long)vblocknr,
 | |
| 		       (unsigned long long)le64_to_cpu(entry->de_start),
 | |
| 		       (unsigned long long)le64_to_cpu(entry->de_end));
 | |
| 		kunmap_atomic(kaddr, KM_USER0);
 | |
| 		brelse(entry_bh);
 | |
| 		return -EINVAL;
 | |
| 	}
 | |
| 	WARN_ON(blocknr == 0);
 | |
| 	entry->de_blocknr = cpu_to_le64(blocknr);
 | |
| 	kunmap_atomic(kaddr, KM_USER0);
 | |
| 
 | |
| 	nilfs_mdt_mark_buffer_dirty(entry_bh);
 | |
| 	nilfs_mdt_mark_dirty(dat);
 | |
| 
 | |
| 	brelse(entry_bh);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * nilfs_dat_translate - translate a virtual block number to a block number
 | |
|  * @dat: DAT file inode
 | |
|  * @vblocknr: virtual block number
 | |
|  * @blocknrp: pointer to a block number
 | |
|  *
 | |
|  * Description: nilfs_dat_translate() maps the virtual block number @vblocknr
 | |
|  * to the corresponding block number.
 | |
|  *
 | |
|  * Return Value: On success, 0 is returned and the block number associated
 | |
|  * with @vblocknr is stored in the place pointed by @blocknrp. On error, one
 | |
|  * of the following negative error codes is returned.
 | |
|  *
 | |
|  * %-EIO - I/O error.
 | |
|  *
 | |
|  * %-ENOMEM - Insufficient amount of memory available.
 | |
|  *
 | |
|  * %-ENOENT - A block number associated with @vblocknr does not exist.
 | |
|  */
 | |
| int nilfs_dat_translate(struct inode *dat, __u64 vblocknr, sector_t *blocknrp)
 | |
| {
 | |
| 	struct buffer_head *entry_bh;
 | |
| 	struct nilfs_dat_entry *entry;
 | |
| 	sector_t blocknr;
 | |
| 	void *kaddr;
 | |
| 	int ret;
 | |
| 
 | |
| 	ret = nilfs_palloc_get_entry_block(dat, vblocknr, 0, &entry_bh);
 | |
| 	if (ret < 0)
 | |
| 		return ret;
 | |
| 
 | |
| 	kaddr = kmap_atomic(entry_bh->b_page, KM_USER0);
 | |
| 	entry = nilfs_palloc_block_get_entry(dat, vblocknr, entry_bh, kaddr);
 | |
| 	blocknr = le64_to_cpu(entry->de_blocknr);
 | |
| 	if (blocknr == 0) {
 | |
| 		ret = -ENOENT;
 | |
| 		goto out;
 | |
| 	}
 | |
| 	if (blocknrp != NULL)
 | |
| 		*blocknrp = blocknr;
 | |
| 
 | |
|  out:
 | |
| 	kunmap_atomic(kaddr, KM_USER0);
 | |
| 	brelse(entry_bh);
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| ssize_t nilfs_dat_get_vinfo(struct inode *dat, struct nilfs_vinfo *vinfo,
 | |
| 			    size_t nvi)
 | |
| {
 | |
| 	struct buffer_head *entry_bh;
 | |
| 	struct nilfs_dat_entry *entry;
 | |
| 	__u64 first, last;
 | |
| 	void *kaddr;
 | |
| 	unsigned long entries_per_block = NILFS_MDT(dat)->mi_entries_per_block;
 | |
| 	int i, j, n, ret;
 | |
| 
 | |
| 	for (i = 0; i < nvi; i += n) {
 | |
| 		ret = nilfs_palloc_get_entry_block(dat, vinfo[i].vi_vblocknr,
 | |
| 						   0, &entry_bh);
 | |
| 		if (ret < 0)
 | |
| 			return ret;
 | |
| 		kaddr = kmap_atomic(entry_bh->b_page, KM_USER0);
 | |
| 		/* last virtual block number in this block */
 | |
| 		first = vinfo[i].vi_vblocknr;
 | |
| 		do_div(first, entries_per_block);
 | |
| 		first *= entries_per_block;
 | |
| 		last = first + entries_per_block - 1;
 | |
| 		for (j = i, n = 0;
 | |
| 		     j < nvi && vinfo[j].vi_vblocknr >= first &&
 | |
| 			     vinfo[j].vi_vblocknr <= last;
 | |
| 		     j++, n++) {
 | |
| 			entry = nilfs_palloc_block_get_entry(
 | |
| 				dat, vinfo[j].vi_vblocknr, entry_bh, kaddr);
 | |
| 			vinfo[j].vi_start = le64_to_cpu(entry->de_start);
 | |
| 			vinfo[j].vi_end = le64_to_cpu(entry->de_end);
 | |
| 			vinfo[j].vi_blocknr = le64_to_cpu(entry->de_blocknr);
 | |
| 		}
 | |
| 		kunmap_atomic(kaddr, KM_USER0);
 | |
| 		brelse(entry_bh);
 | |
| 	}
 | |
| 
 | |
| 	return nvi;
 | |
| }
 |