232 lines
		
	
	
		
			5.5 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			232 lines
		
	
	
		
			5.5 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
/*
 | 
						|
 * Copyright (C) 2008 Oracle.  All rights reserved.
 | 
						|
 *
 | 
						|
 * This program is free software; you can redistribute it and/or
 | 
						|
 * modify it under the terms of the GNU General Public
 | 
						|
 * License v2 as published by the Free Software Foundation.
 | 
						|
 *
 | 
						|
 * This program is distributed in the hope that it will be useful,
 | 
						|
 * but WITHOUT ANY WARRANTY; without even the implied warranty of
 | 
						|
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 | 
						|
 * General Public License for more details.
 | 
						|
 *
 | 
						|
 * You should have received a copy of the GNU General Public
 | 
						|
 * License along with this program; if not, write to the
 | 
						|
 * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
 | 
						|
 * Boston, MA 021110-1307, USA.
 | 
						|
 */
 | 
						|
 | 
						|
#include <linux/sched.h>
 | 
						|
#include <linux/sort.h>
 | 
						|
#include "ctree.h"
 | 
						|
#include "ref-cache.h"
 | 
						|
#include "transaction.h"
 | 
						|
 | 
						|
/*
 | 
						|
 * leaf refs are used to cache the information about which extents
 | 
						|
 * a given leaf has references on.  This allows us to process that leaf
 | 
						|
 * in btrfs_drop_snapshot without needing to read it back from disk.
 | 
						|
 */
 | 
						|
 | 
						|
/*
 | 
						|
 * kmalloc a leaf reference struct and update the counters for the
 | 
						|
 * total ref cache size
 | 
						|
 */
 | 
						|
struct btrfs_leaf_ref *btrfs_alloc_leaf_ref(struct btrfs_root *root,
 | 
						|
					    int nr_extents)
 | 
						|
{
 | 
						|
	struct btrfs_leaf_ref *ref;
 | 
						|
	size_t size = btrfs_leaf_ref_size(nr_extents);
 | 
						|
 | 
						|
	ref = kmalloc(size, GFP_NOFS);
 | 
						|
	if (ref) {
 | 
						|
		spin_lock(&root->fs_info->ref_cache_lock);
 | 
						|
		root->fs_info->total_ref_cache_size += size;
 | 
						|
		spin_unlock(&root->fs_info->ref_cache_lock);
 | 
						|
 | 
						|
		memset(ref, 0, sizeof(*ref));
 | 
						|
		atomic_set(&ref->usage, 1);
 | 
						|
		INIT_LIST_HEAD(&ref->list);
 | 
						|
	}
 | 
						|
	return ref;
 | 
						|
}
 | 
						|
 | 
						|
/*
 | 
						|
 * free a leaf reference struct and update the counters for the
 | 
						|
 * total ref cache size
 | 
						|
 */
 | 
						|
void btrfs_free_leaf_ref(struct btrfs_root *root, struct btrfs_leaf_ref *ref)
 | 
						|
{
 | 
						|
	if (!ref)
 | 
						|
		return;
 | 
						|
	WARN_ON(atomic_read(&ref->usage) == 0);
 | 
						|
	if (atomic_dec_and_test(&ref->usage)) {
 | 
						|
		size_t size = btrfs_leaf_ref_size(ref->nritems);
 | 
						|
 | 
						|
		BUG_ON(ref->in_tree);
 | 
						|
		kfree(ref);
 | 
						|
 | 
						|
		spin_lock(&root->fs_info->ref_cache_lock);
 | 
						|
		root->fs_info->total_ref_cache_size -= size;
 | 
						|
		spin_unlock(&root->fs_info->ref_cache_lock);
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
static struct rb_node *tree_insert(struct rb_root *root, u64 bytenr,
 | 
						|
				   struct rb_node *node)
 | 
						|
{
 | 
						|
	struct rb_node **p = &root->rb_node;
 | 
						|
	struct rb_node *parent = NULL;
 | 
						|
	struct btrfs_leaf_ref *entry;
 | 
						|
 | 
						|
	while (*p) {
 | 
						|
		parent = *p;
 | 
						|
		entry = rb_entry(parent, struct btrfs_leaf_ref, rb_node);
 | 
						|
 | 
						|
		if (bytenr < entry->bytenr)
 | 
						|
			p = &(*p)->rb_left;
 | 
						|
		else if (bytenr > entry->bytenr)
 | 
						|
			p = &(*p)->rb_right;
 | 
						|
		else
 | 
						|
			return parent;
 | 
						|
	}
 | 
						|
 | 
						|
	entry = rb_entry(node, struct btrfs_leaf_ref, rb_node);
 | 
						|
	rb_link_node(node, parent, p);
 | 
						|
	rb_insert_color(node, root);
 | 
						|
	return NULL;
 | 
						|
}
 | 
						|
 | 
						|
static struct rb_node *tree_search(struct rb_root *root, u64 bytenr)
 | 
						|
{
 | 
						|
	struct rb_node *n = root->rb_node;
 | 
						|
	struct btrfs_leaf_ref *entry;
 | 
						|
 | 
						|
	while (n) {
 | 
						|
		entry = rb_entry(n, struct btrfs_leaf_ref, rb_node);
 | 
						|
		WARN_ON(!entry->in_tree);
 | 
						|
 | 
						|
		if (bytenr < entry->bytenr)
 | 
						|
			n = n->rb_left;
 | 
						|
		else if (bytenr > entry->bytenr)
 | 
						|
			n = n->rb_right;
 | 
						|
		else
 | 
						|
			return n;
 | 
						|
	}
 | 
						|
	return NULL;
 | 
						|
}
 | 
						|
 | 
						|
int btrfs_remove_leaf_refs(struct btrfs_root *root, u64 max_root_gen,
 | 
						|
			   int shared)
 | 
						|
{
 | 
						|
	struct btrfs_leaf_ref *ref = NULL;
 | 
						|
	struct btrfs_leaf_ref_tree *tree = root->ref_tree;
 | 
						|
 | 
						|
	if (shared)
 | 
						|
		tree = &root->fs_info->shared_ref_tree;
 | 
						|
	if (!tree)
 | 
						|
		return 0;
 | 
						|
 | 
						|
	spin_lock(&tree->lock);
 | 
						|
	while (!list_empty(&tree->list)) {
 | 
						|
		ref = list_entry(tree->list.next, struct btrfs_leaf_ref, list);
 | 
						|
		BUG_ON(ref->tree != tree);
 | 
						|
		if (ref->root_gen > max_root_gen)
 | 
						|
			break;
 | 
						|
		if (!xchg(&ref->in_tree, 0)) {
 | 
						|
			cond_resched_lock(&tree->lock);
 | 
						|
			continue;
 | 
						|
		}
 | 
						|
 | 
						|
		rb_erase(&ref->rb_node, &tree->root);
 | 
						|
		list_del_init(&ref->list);
 | 
						|
 | 
						|
		spin_unlock(&tree->lock);
 | 
						|
		btrfs_free_leaf_ref(root, ref);
 | 
						|
		cond_resched();
 | 
						|
		spin_lock(&tree->lock);
 | 
						|
	}
 | 
						|
	spin_unlock(&tree->lock);
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
/*
 | 
						|
 * find the leaf ref for a given extent.  This returns the ref struct with
 | 
						|
 * a usage reference incremented
 | 
						|
 */
 | 
						|
struct btrfs_leaf_ref *btrfs_lookup_leaf_ref(struct btrfs_root *root,
 | 
						|
					     u64 bytenr)
 | 
						|
{
 | 
						|
	struct rb_node *rb;
 | 
						|
	struct btrfs_leaf_ref *ref = NULL;
 | 
						|
	struct btrfs_leaf_ref_tree *tree = root->ref_tree;
 | 
						|
again:
 | 
						|
	if (tree) {
 | 
						|
		spin_lock(&tree->lock);
 | 
						|
		rb = tree_search(&tree->root, bytenr);
 | 
						|
		if (rb)
 | 
						|
			ref = rb_entry(rb, struct btrfs_leaf_ref, rb_node);
 | 
						|
		if (ref)
 | 
						|
			atomic_inc(&ref->usage);
 | 
						|
		spin_unlock(&tree->lock);
 | 
						|
		if (ref)
 | 
						|
			return ref;
 | 
						|
	}
 | 
						|
	if (tree != &root->fs_info->shared_ref_tree) {
 | 
						|
		tree = &root->fs_info->shared_ref_tree;
 | 
						|
		goto again;
 | 
						|
	}
 | 
						|
	return NULL;
 | 
						|
}
 | 
						|
 | 
						|
/*
 | 
						|
 * add a fully filled in leaf ref struct
 | 
						|
 * remove all the refs older than a given root generation
 | 
						|
 */
 | 
						|
int btrfs_add_leaf_ref(struct btrfs_root *root, struct btrfs_leaf_ref *ref,
 | 
						|
		       int shared)
 | 
						|
{
 | 
						|
	int ret = 0;
 | 
						|
	struct rb_node *rb;
 | 
						|
	struct btrfs_leaf_ref_tree *tree = root->ref_tree;
 | 
						|
 | 
						|
	if (shared)
 | 
						|
		tree = &root->fs_info->shared_ref_tree;
 | 
						|
 | 
						|
	spin_lock(&tree->lock);
 | 
						|
	rb = tree_insert(&tree->root, ref->bytenr, &ref->rb_node);
 | 
						|
	if (rb) {
 | 
						|
		ret = -EEXIST;
 | 
						|
	} else {
 | 
						|
		atomic_inc(&ref->usage);
 | 
						|
		ref->tree = tree;
 | 
						|
		ref->in_tree = 1;
 | 
						|
		list_add_tail(&ref->list, &tree->list);
 | 
						|
	}
 | 
						|
	spin_unlock(&tree->lock);
 | 
						|
	return ret;
 | 
						|
}
 | 
						|
 | 
						|
/*
 | 
						|
 * remove a single leaf ref from the tree.  This drops the ref held by the tree
 | 
						|
 * only
 | 
						|
 */
 | 
						|
int btrfs_remove_leaf_ref(struct btrfs_root *root, struct btrfs_leaf_ref *ref)
 | 
						|
{
 | 
						|
	struct btrfs_leaf_ref_tree *tree;
 | 
						|
 | 
						|
	if (!xchg(&ref->in_tree, 0))
 | 
						|
		return 0;
 | 
						|
 | 
						|
	tree = ref->tree;
 | 
						|
	spin_lock(&tree->lock);
 | 
						|
 | 
						|
	rb_erase(&ref->rb_node, &tree->root);
 | 
						|
	list_del_init(&ref->list);
 | 
						|
 | 
						|
	spin_unlock(&tree->lock);
 | 
						|
 | 
						|
	btrfs_free_leaf_ref(root, ref);
 | 
						|
	return 0;
 | 
						|
}
 |