mirror of https://github.com/OpenIPC/firmware.git
				
				
				
			
		
			
				
	
	
		
			1718 lines
		
	
	
		
			45 KiB
		
	
	
	
		
			Diff
		
	
	
			
		
		
	
	
			1718 lines
		
	
	
		
			45 KiB
		
	
	
	
		
			Diff
		
	
	
| diff -drupN a/fs/yaffs2/yaffs_yaffs2.c b/fs/yaffs2/yaffs_yaffs2.c
 | |
| --- a/fs/yaffs2/yaffs_yaffs2.c	1970-01-01 03:00:00.000000000 +0300
 | |
| +++ b/fs/yaffs2/yaffs_yaffs2.c	2022-06-09 05:02:35.000000000 +0300
 | |
| @@ -0,0 +1,1713 @@
 | |
| +/*
 | |
| + * YAFFS: Yet Another Flash File System. A NAND-flash specific file system.
 | |
| + *
 | |
| + * Copyright (C) 2002-2018 Aleph One Ltd.
 | |
| + *
 | |
| + * Created by Charles Manning <charles@aleph1.co.uk>
 | |
| + *
 | |
| + * This program is free software; you can redistribute it and/or modify
 | |
| + * it under the terms of the GNU General Public License version 2 as
 | |
| + * published by the Free Software Foundation.
 | |
| + */
 | |
| +
 | |
| +#include "yaffs_guts.h"
 | |
| +#include "yaffs_trace.h"
 | |
| +#include "yaffs_yaffs2.h"
 | |
| +#include "yaffs_checkptrw.h"
 | |
| +#include "yaffs_bitmap.h"
 | |
| +#include "yaffs_nand.h"
 | |
| +#include "yaffs_getblockinfo.h"
 | |
| +#include "yaffs_verify.h"
 | |
| +#include "yaffs_attribs.h"
 | |
| +#include "yaffs_summary.h"
 | |
| +#include "yaffs_endian.h"
 | |
| +
 | |
| +/*
 | |
| + * Checkpoints are really no benefit on very small partitions.
 | |
| + *
 | |
| + * To save space on small partitions don't bother with checkpoints unless
 | |
| + * the partition is at least this big.
 | |
| + */
 | |
| +#define YAFFS_CHECKPOINT_MIN_BLOCKS 60
 | |
| +#define YAFFS_SMALL_HOLE_THRESHOLD 4
 | |
| +
 | |
| +/*
 | |
| + * Oldest Dirty Sequence Number handling.
 | |
| + */
 | |
| +
 | |
| +/* yaffs_calc_oldest_dirty_seq()
 | |
| + * yaffs2_find_oldest_dirty_seq()
 | |
| + * Calculate the oldest dirty sequence number if we don't know it.
 | |
| + */
 | |
| +void yaffs_calc_oldest_dirty_seq(struct yaffs_dev *dev)
 | |
| +{
 | |
| +	u32 i;
 | |
| +	unsigned seq;
 | |
| +	unsigned block_no = 0;
 | |
| +	struct yaffs_block_info *b;
 | |
| +
 | |
| +	if (!dev->param.is_yaffs2)
 | |
| +		return;
 | |
| +
 | |
| +	/* Find the oldest dirty sequence number. */
 | |
| +	seq = dev->seq_number + 1;
 | |
| +	b = dev->block_info;
 | |
| +	for (i = dev->internal_start_block; i <= dev->internal_end_block; i++) {
 | |
| +		if (b->block_state == YAFFS_BLOCK_STATE_FULL &&
 | |
| +		    (u32)(b->pages_in_use - b->soft_del_pages) <
 | |
| +		    dev->param.chunks_per_block &&
 | |
| +		    b->seq_number < seq) {
 | |
| +			seq = b->seq_number;
 | |
| +			block_no = i;
 | |
| +		}
 | |
| +		b++;
 | |
| +	}
 | |
| +
 | |
| +	if (block_no) {
 | |
| +		dev->oldest_dirty_seq = seq;
 | |
| +		dev->oldest_dirty_block = block_no;
 | |
| +	}
 | |
| +}
 | |
| +
 | |
| +void yaffs2_find_oldest_dirty_seq(struct yaffs_dev *dev)
 | |
| +{
 | |
| +	if (!dev->param.is_yaffs2)
 | |
| +		return;
 | |
| +
 | |
| +	if (!dev->oldest_dirty_seq)
 | |
| +		yaffs_calc_oldest_dirty_seq(dev);
 | |
| +}
 | |
| +
 | |
| +/*
 | |
| + * yaffs_clear_oldest_dirty_seq()
 | |
| + * Called when a block is erased or marked bad. (ie. when its seq_number
 | |
| + * becomes invalid). If the value matches the oldest then we clear
 | |
| + * dev->oldest_dirty_seq to force its recomputation.
 | |
| + */
 | |
| +void yaffs2_clear_oldest_dirty_seq(struct yaffs_dev *dev,
 | |
| +				   struct yaffs_block_info *bi)
 | |
| +{
 | |
| +
 | |
| +	if (!dev->param.is_yaffs2)
 | |
| +		return;
 | |
| +
 | |
| +	if (!bi || bi->seq_number == dev->oldest_dirty_seq) {
 | |
| +		dev->oldest_dirty_seq = 0;
 | |
| +		dev->oldest_dirty_block = 0;
 | |
| +	}
 | |
| +}
 | |
| +
 | |
| +/*
 | |
| + * yaffs2_update_oldest_dirty_seq()
 | |
| + * Update the oldest dirty sequence number whenever we dirty a block.
 | |
| + * Only do this if the oldest_dirty_seq is actually being tracked.
 | |
| + */
 | |
| +void yaffs2_update_oldest_dirty_seq(struct yaffs_dev *dev, unsigned block_no,
 | |
| +				    struct yaffs_block_info *bi)
 | |
| +{
 | |
| +	if (!dev->param.is_yaffs2)
 | |
| +		return;
 | |
| +
 | |
| +	if (dev->oldest_dirty_seq) {
 | |
| +		if (dev->oldest_dirty_seq > bi->seq_number) {
 | |
| +			dev->oldest_dirty_seq = bi->seq_number;
 | |
| +			dev->oldest_dirty_block = block_no;
 | |
| +		}
 | |
| +	}
 | |
| +}
 | |
| +
 | |
| +int yaffs_block_ok_for_gc(struct yaffs_dev *dev, struct yaffs_block_info *bi)
 | |
| +{
 | |
| +
 | |
| +	if (!dev->param.is_yaffs2)
 | |
| +		return 1;	/* disqualification only applies to yaffs2. */
 | |
| +
 | |
| +	if (!bi->has_shrink_hdr)
 | |
| +		return 1;	/* can gc */
 | |
| +
 | |
| +	yaffs2_find_oldest_dirty_seq(dev);
 | |
| +
 | |
| +	/* Can't do gc of this block if there are any blocks older than this
 | |
| +	 * one that have discarded pages.
 | |
| +	 */
 | |
| +	return (bi->seq_number <= dev->oldest_dirty_seq);
 | |
| +}
 | |
| +
 | |
| +/*
 | |
| + * yaffs2_find_refresh_block()
 | |
| + * periodically finds the oldest full block by sequence number for refreshing.
 | |
| + * Only for yaffs2.
 | |
| + */
 | |
| +u32 yaffs2_find_refresh_block(struct yaffs_dev *dev)
 | |
| +{
 | |
| +	u32 b;
 | |
| +	u32 oldest = 0;
 | |
| +	u32 oldest_seq = 0;
 | |
| +	struct yaffs_block_info *bi;
 | |
| +
 | |
| +	if (!dev->param.is_yaffs2)
 | |
| +		return oldest;
 | |
| +
 | |
| +	/*
 | |
| +	 * If refresh period < 10 then refreshing is disabled.
 | |
| +	 */
 | |
| +	if (dev->param.refresh_period < 10)
 | |
| +		return oldest;
 | |
| +
 | |
| +	/*
 | |
| +	 * Fix broken values.
 | |
| +	 */
 | |
| +	if (dev->refresh_skip > dev->param.refresh_period)
 | |
| +		dev->refresh_skip = dev->param.refresh_period;
 | |
| +
 | |
| +	if (dev->refresh_skip > 0)
 | |
| +		return oldest;
 | |
| +
 | |
| +	/*
 | |
| +	 * Refresh skip is now zero.
 | |
| +	 * We'll do a refresh this time around....
 | |
| +	 * Update the refresh skip and find the oldest block.
 | |
| +	 */
 | |
| +	dev->refresh_skip = dev->param.refresh_period;
 | |
| +	dev->refresh_count++;
 | |
| +	bi = dev->block_info;
 | |
| +	for (b = dev->internal_start_block; b <= dev->internal_end_block; b++) {
 | |
| +
 | |
| +		if (bi->block_state == YAFFS_BLOCK_STATE_FULL) {
 | |
| +
 | |
| +			if (oldest < 1 || bi->seq_number < oldest_seq) {
 | |
| +				oldest = b;
 | |
| +				oldest_seq = bi->seq_number;
 | |
| +			}
 | |
| +		}
 | |
| +		bi++;
 | |
| +	}
 | |
| +
 | |
| +	if (oldest > 0) {
 | |
| +		yaffs_trace(YAFFS_TRACE_GC,
 | |
| +			"GC refresh count %d selected block %d with seq_number %d",
 | |
| +			dev->refresh_count, oldest, oldest_seq);
 | |
| +	}
 | |
| +
 | |
| +	return oldest;
 | |
| +}
 | |
| +
 | |
| +int yaffs2_checkpt_required(struct yaffs_dev *dev)
 | |
| +{
 | |
| +	int nblocks;
 | |
| +
 | |
| +	if (!dev->param.is_yaffs2)
 | |
| +		return 0;
 | |
| +
 | |
| +	nblocks = dev->internal_end_block - dev->internal_start_block + 1;
 | |
| +
 | |
| +	return !dev->param.skip_checkpt_wr &&
 | |
| +	    !dev->read_only && (nblocks >= YAFFS_CHECKPOINT_MIN_BLOCKS);
 | |
| +}
 | |
| +
 | |
| +int yaffs_calc_checkpt_blocks_required(struct yaffs_dev *dev)
 | |
| +{
 | |
| +	int retval;
 | |
| +	int n_bytes = 0;
 | |
| +	int n_blocks;
 | |
| +	int dev_blocks;
 | |
| +
 | |
| +	if (!dev->param.is_yaffs2)
 | |
| +		return 0;
 | |
| +
 | |
| +	if (!dev->checkpoint_blocks_required && yaffs2_checkpt_required(dev)) {
 | |
| +		/* Not a valid value so recalculate */
 | |
| +		dev_blocks = dev->param.end_block - dev->param.start_block + 1;
 | |
| +		n_bytes += sizeof(struct yaffs_checkpt_validity);
 | |
| +		n_bytes += sizeof(struct yaffs_checkpt_dev);
 | |
| +		n_bytes += dev_blocks * sizeof(struct yaffs_block_info);
 | |
| +		n_bytes += dev_blocks * dev->chunk_bit_stride;
 | |
| +		n_bytes +=
 | |
| +		    (sizeof(struct yaffs_checkpt_obj) + sizeof(u32)) *
 | |
| +		    dev->n_obj;
 | |
| +		n_bytes += (dev->tnode_size + sizeof(u32)) * dev->n_tnodes;
 | |
| +		n_bytes += sizeof(struct yaffs_checkpt_validity);
 | |
| +		n_bytes += sizeof(u32);	/* checksum */
 | |
| +
 | |
| +		/* Round up and add 2 blocks to allow for some bad blocks,
 | |
| +		 * so add 3 */
 | |
| +
 | |
| +		n_blocks =
 | |
| +		    (n_bytes /
 | |
| +		     (dev->data_bytes_per_chunk *
 | |
| +		      dev->param.chunks_per_block)) + 3;
 | |
| +
 | |
| +		dev->checkpoint_blocks_required = n_blocks;
 | |
| +	}
 | |
| +
 | |
| +	retval = dev->checkpoint_blocks_required - dev->blocks_in_checkpt;
 | |
| +	if (retval < 0)
 | |
| +		retval = 0;
 | |
| +	return retval;
 | |
| +}
 | |
| +
 | |
| +/*--------------------- Checkpointing --------------------*/
 | |
| +
 | |
| +static void yaffs2_do_endian_validity_marker(struct yaffs_dev *dev,
 | |
| +					     struct yaffs_checkpt_validity *v)
 | |
| +{
 | |
| +
 | |
| +	if (!dev->swap_endian)
 | |
| +		return;
 | |
| +	v->struct_type = swap_s32(v->struct_type);
 | |
| +	v->magic = swap_u32(v->magic);
 | |
| +	v->version = swap_u32(v->version);
 | |
| +	v->head = swap_u32(v->head);
 | |
| +}
 | |
| +
 | |
| +static int yaffs2_wr_checkpt_validity_marker(struct yaffs_dev *dev, int head)
 | |
| +{
 | |
| +	struct yaffs_checkpt_validity cp;
 | |
| +
 | |
| +	memset(&cp, 0, sizeof(cp));
 | |
| +
 | |
| +	cp.struct_type = sizeof(cp);
 | |
| +	cp.magic = YAFFS_MAGIC;
 | |
| +	cp.version = YAFFS_CHECKPOINT_VERSION;
 | |
| +	cp.head = (head) ? 1 : 0;
 | |
| +
 | |
| +	yaffs2_do_endian_validity_marker(dev, &cp);
 | |
| +
 | |
| +	return (yaffs2_checkpt_wr(dev, &cp, sizeof(cp)) == sizeof(cp)) ? 1 : 0;
 | |
| +}
 | |
| +
 | |
| +static int yaffs2_rd_checkpt_validity_marker(struct yaffs_dev *dev, int head)
 | |
| +{
 | |
| +	struct yaffs_checkpt_validity cp;
 | |
| +	int ok;
 | |
| +
 | |
| +	ok = (yaffs2_checkpt_rd(dev, &cp, sizeof(cp)) == sizeof(cp));
 | |
| +	yaffs2_do_endian_validity_marker(dev, &cp);
 | |
| +
 | |
| +	if (ok)
 | |
| +		ok = (cp.struct_type == sizeof(cp)) &&
 | |
| +		    (cp.magic == YAFFS_MAGIC) &&
 | |
| +		    (cp.version == YAFFS_CHECKPOINT_VERSION) &&
 | |
| +		    (cp.head == ((head) ? 1 : 0));
 | |
| +	return ok ? 1 : 0;
 | |
| +}
 | |
| +
 | |
| +static void yaffs2_dev_to_checkpt_dev(struct yaffs_checkpt_dev *cp,
 | |
| +				      struct yaffs_dev *dev)
 | |
| +{
 | |
| +	cp->struct_type = sizeof(*cp);
 | |
| +
 | |
| +	cp->n_erased_blocks = dev->n_erased_blocks;
 | |
| +	cp->alloc_block = dev->alloc_block;
 | |
| +	cp->alloc_page = dev->alloc_page;
 | |
| +	cp->n_free_chunks = dev->n_free_chunks;
 | |
| +
 | |
| +	cp->n_deleted_files = dev->n_deleted_files;
 | |
| +	cp->n_unlinked_files = dev->n_unlinked_files;
 | |
| +	cp->n_bg_deletions = dev->n_bg_deletions;
 | |
| +	cp->seq_number = dev->seq_number;
 | |
| +
 | |
| +}
 | |
| +
 | |
| +static void yaffs_checkpt_dev_to_dev(struct yaffs_dev *dev,
 | |
| +				     struct yaffs_checkpt_dev *cp)
 | |
| +{
 | |
| +	dev->n_erased_blocks = cp->n_erased_blocks;
 | |
| +	dev->alloc_block = cp->alloc_block;
 | |
| +	dev->alloc_page = cp->alloc_page;
 | |
| +	dev->n_free_chunks = cp->n_free_chunks;
 | |
| +
 | |
| +	dev->n_deleted_files = cp->n_deleted_files;
 | |
| +	dev->n_unlinked_files = cp->n_unlinked_files;
 | |
| +	dev->n_bg_deletions = cp->n_bg_deletions;
 | |
| +	dev->seq_number = cp->seq_number;
 | |
| +}
 | |
| +
 | |
| +static void yaffs2_do_endian_checkpt_dev(struct yaffs_dev *dev,
 | |
| +				     struct yaffs_checkpt_dev *cp)
 | |
| +{
 | |
| +	if (!dev->swap_endian)
 | |
| +		return;
 | |
| +	cp->struct_type = swap_s32(cp->struct_type);
 | |
| +	cp->n_erased_blocks = swap_s32(cp->n_erased_blocks);
 | |
| +	cp->alloc_block = swap_s32(cp->alloc_block);
 | |
| +	cp->alloc_page = swap_u32(cp->alloc_page);
 | |
| +	cp->n_free_chunks = swap_s32(cp->n_free_chunks);
 | |
| +	cp->n_deleted_files = swap_s32(cp->n_deleted_files);
 | |
| +	cp->n_unlinked_files = swap_s32(cp->n_unlinked_files);
 | |
| +	cp->n_bg_deletions = swap_s32(cp->n_bg_deletions);
 | |
| +}
 | |
| +
 | |
| +static int yaffs2_wr_checkpt_dev(struct yaffs_dev *dev)
 | |
| +{
 | |
| +	struct yaffs_checkpt_dev cp;
 | |
| +	u32 n_bytes;
 | |
| +	u32 n_blocks = dev->internal_end_block - dev->internal_start_block + 1;
 | |
| +	int ok;
 | |
| +	u32 i;
 | |
| +	union yaffs_block_info_union bu;
 | |
| +
 | |
| +	/* Write device runtime values */
 | |
| +	yaffs2_dev_to_checkpt_dev(&cp, dev);
 | |
| +	yaffs2_do_endian_checkpt_dev(dev, &cp);
 | |
| +
 | |
| +	ok = (yaffs2_checkpt_wr(dev, &cp, sizeof(cp)) == sizeof(cp));
 | |
| +	if (!ok)
 | |
| +		return 0;
 | |
| +
 | |
| +	/* Write block info. */
 | |
| +	if (!dev->swap_endian) {
 | |
| +		n_bytes = n_blocks * sizeof(struct yaffs_block_info);
 | |
| +		ok = (yaffs2_checkpt_wr(dev, dev->block_info, n_bytes) ==
 | |
| +			(int)n_bytes);
 | |
| +	} else {
 | |
| +		/*
 | |
| +		 * Need to swap the endianisms. We can't do this in place
 | |
| +		 * since that would damage live data,
 | |
| +		 * so write one block info at a time using a copy.
 | |
| +		 */
 | |
| +		for (i = 0; i < n_blocks && ok; i++) {
 | |
| +			bu.bi = dev->block_info[i];
 | |
| +			bu.as_u32[0] = swap_u32(bu.as_u32[0]);
 | |
| +			bu.as_u32[1] = swap_u32(bu.as_u32[1]);
 | |
| +			ok = (yaffs2_checkpt_wr(dev, &bu, sizeof(bu)) == sizeof(bu));
 | |
| +		}
 | |
| +	}
 | |
| +
 | |
| +	if (!ok)
 | |
| +		return 0;
 | |
| +
 | |
| +	/*
 | |
| +	 * Write chunk bits. Chunk bits are in bytes so
 | |
| +	 * no endian conversion is needed.
 | |
| +	 */
 | |
| +	n_bytes = n_blocks * dev->chunk_bit_stride;
 | |
| +	ok = (yaffs2_checkpt_wr(dev, dev->chunk_bits, n_bytes) ==
 | |
| +		(int)n_bytes);
 | |
| +
 | |
| +	return ok ? 1 : 0;
 | |
| +}
 | |
| +
 | |
| +static int yaffs2_rd_checkpt_dev(struct yaffs_dev *dev)
 | |
| +{
 | |
| +	struct yaffs_checkpt_dev cp;
 | |
| +	u32 n_bytes;
 | |
| +	u32 n_blocks =
 | |
| +	    (dev->internal_end_block - dev->internal_start_block + 1);
 | |
| +	int ok;
 | |
| +
 | |
| +	ok = (yaffs2_checkpt_rd(dev, &cp, sizeof(cp)) == sizeof(cp));
 | |
| +	if (!ok)
 | |
| +		return 0;
 | |
| +	yaffs2_do_endian_checkpt_dev(dev, &cp);
 | |
| +
 | |
| +	if (cp.struct_type != sizeof(cp))
 | |
| +		return 0;
 | |
| +
 | |
| +	yaffs_checkpt_dev_to_dev(dev, &cp);
 | |
| +
 | |
| +	n_bytes = n_blocks * sizeof(struct yaffs_block_info);
 | |
| +
 | |
| +	ok = (yaffs2_checkpt_rd(dev, dev->block_info, n_bytes) ==
 | |
| +		(int)n_bytes);
 | |
| +
 | |
| +	if (!ok)
 | |
| +		return 0;
 | |
| +
 | |
| +	if (dev->swap_endian) {
 | |
| +		/* The block info can just be handled as a list of u32s. */
 | |
| +		u32 *as_u32 = (u32 *) dev->block_info;
 | |
| +		u32 n_u32s = n_bytes/sizeof(u32);
 | |
| +		u32 i;
 | |
| +
 | |
| +		for (i=0; i < n_u32s; i++)
 | |
| +			as_u32[i] = swap_u32(as_u32[i]);
 | |
| +	}
 | |
| +
 | |
| +	n_bytes = n_blocks * dev->chunk_bit_stride;
 | |
| +
 | |
| +	ok = (yaffs2_checkpt_rd(dev, dev->chunk_bits, n_bytes) ==
 | |
| +		(int)n_bytes);
 | |
| +
 | |
| +
 | |
| +	return ok ? 1 : 0;
 | |
| +}
 | |
| +
 | |
| +
 | |
| +static void yaffs2_checkpt_obj_bit_assign(struct yaffs_checkpt_obj *cp,
 | |
| +					  int bit_offset,
 | |
| +					  int bit_width,
 | |
| +					  u32 value)
 | |
| +{
 | |
| +	u32 and_mask;
 | |
| +
 | |
| +	and_mask = ((1<<bit_width)-1) << bit_offset;
 | |
| +
 | |
| +	cp->bit_field &= ~and_mask;
 | |
| +	cp->bit_field |= ((value << bit_offset) & and_mask);
 | |
| +}
 | |
| +
 | |
| +static u32 yaffs2_checkpt_obj_bit_get(struct yaffs_checkpt_obj *cp,
 | |
| +				      int bit_offset,
 | |
| +				      int bit_width)
 | |
| +{
 | |
| +	u32 and_mask;
 | |
| +
 | |
| +	and_mask = ((1<<bit_width)-1);
 | |
| +
 | |
| +	return (cp->bit_field >> bit_offset) & and_mask;
 | |
| +}
 | |
| +
 | |
| +static void yaffs2_obj_checkpt_obj(struct yaffs_checkpt_obj *cp,
 | |
| +				   struct yaffs_obj *obj)
 | |
| +{
 | |
| +	cp->obj_id = obj->obj_id;
 | |
| +	cp->parent_id = (obj->parent) ? obj->parent->obj_id : 0;
 | |
| +	cp->hdr_chunk = obj->hdr_chunk;
 | |
| +
 | |
| +	yaffs2_checkpt_obj_bit_assign(cp, CHECKPOINT_VARIANT_BITS, obj->variant_type);
 | |
| +	yaffs2_checkpt_obj_bit_assign(cp, CHECKPOINT_DELETED_BITS, obj->deleted);
 | |
| +	yaffs2_checkpt_obj_bit_assign(cp, CHECKPOINT_SOFT_DEL_BITS, obj->soft_del);
 | |
| +	yaffs2_checkpt_obj_bit_assign(cp, CHECKPOINT_UNLINKED_BITS, obj->unlinked);
 | |
| +	yaffs2_checkpt_obj_bit_assign(cp, CHECKPOINT_FAKE_BITS, obj->fake);
 | |
| +	yaffs2_checkpt_obj_bit_assign(cp, CHECKPOINT_RENAME_ALLOWED_BITS, obj->rename_allowed);
 | |
| +	yaffs2_checkpt_obj_bit_assign(cp, CHECKPOINT_UNLINK_ALLOWED_BITS, obj->unlink_allowed);
 | |
| +	yaffs2_checkpt_obj_bit_assign(cp, CHECKPOINT_SERIAL_BITS, obj->serial);
 | |
| +
 | |
| +	cp->n_data_chunks = obj->n_data_chunks;
 | |
| +
 | |
| +	if (obj->variant_type == YAFFS_OBJECT_TYPE_FILE)
 | |
| +		cp->size_or_equiv_obj = obj->variant.file_variant.file_size;
 | |
| +	else if (obj->variant_type == YAFFS_OBJECT_TYPE_HARDLINK)
 | |
| +		cp->size_or_equiv_obj = obj->variant.hardlink_variant.equiv_id;
 | |
| +}
 | |
| +
 | |
| +static int yaffs2_checkpt_obj_to_obj(struct yaffs_obj *obj,
 | |
| +				     struct yaffs_checkpt_obj *cp)
 | |
| +{
 | |
| +	struct yaffs_obj *parent;
 | |
| +	u32 cp_variant_type = yaffs2_checkpt_obj_bit_get(cp, CHECKPOINT_VARIANT_BITS);
 | |
| +
 | |
| +	if (obj->variant_type != cp_variant_type) {
 | |
| +		yaffs_trace(YAFFS_TRACE_ERROR,
 | |
| +			"Checkpoint read object %d type %d chunk %d does not match existing object type %d",
 | |
| +			cp->obj_id, cp_variant_type, cp->hdr_chunk,
 | |
| +			obj->variant_type);
 | |
| +		return 0;
 | |
| +	}
 | |
| +
 | |
| +	obj->obj_id = cp->obj_id;
 | |
| +
 | |
| +	if (cp->parent_id)
 | |
| +		parent = yaffs_find_or_create_by_number(obj->my_dev,
 | |
| +						cp->parent_id,
 | |
| +						YAFFS_OBJECT_TYPE_DIRECTORY);
 | |
| +	else
 | |
| +		parent = NULL;
 | |
| +
 | |
| +	if (parent) {
 | |
| +		if (parent->variant_type != YAFFS_OBJECT_TYPE_DIRECTORY) {
 | |
| +			yaffs_trace(YAFFS_TRACE_ALWAYS,
 | |
| +				"Checkpoint read object %d parent %d type %d chunk %d Parent type, %d, not directory",
 | |
| +				cp->obj_id, cp->parent_id,
 | |
| +				cp_variant_type, cp->hdr_chunk,
 | |
| +				parent->variant_type);
 | |
| +			return 0;
 | |
| +		}
 | |
| +		yaffs_add_obj_to_dir(parent, obj);
 | |
| +	}
 | |
| +
 | |
| +	obj->hdr_chunk = cp->hdr_chunk;
 | |
| +
 | |
| +	obj->variant_type = yaffs2_checkpt_obj_bit_get(cp, CHECKPOINT_VARIANT_BITS);
 | |
| +	obj->deleted = yaffs2_checkpt_obj_bit_get(cp, CHECKPOINT_DELETED_BITS);
 | |
| +	obj->soft_del = yaffs2_checkpt_obj_bit_get(cp, CHECKPOINT_SOFT_DEL_BITS);
 | |
| +	obj->unlinked = yaffs2_checkpt_obj_bit_get(cp, CHECKPOINT_UNLINKED_BITS);
 | |
| +	obj->fake = yaffs2_checkpt_obj_bit_get(cp, CHECKPOINT_FAKE_BITS);
 | |
| +	obj->rename_allowed = yaffs2_checkpt_obj_bit_get(cp, CHECKPOINT_RENAME_ALLOWED_BITS);
 | |
| +	obj->unlink_allowed = yaffs2_checkpt_obj_bit_get(cp, CHECKPOINT_UNLINK_ALLOWED_BITS);
 | |
| +	obj->serial = yaffs2_checkpt_obj_bit_get(cp, CHECKPOINT_SERIAL_BITS);
 | |
| +
 | |
| +	obj->n_data_chunks = cp->n_data_chunks;
 | |
| +
 | |
| +	if (obj->variant_type == YAFFS_OBJECT_TYPE_FILE) {
 | |
| +		obj->variant.file_variant.file_size = cp->size_or_equiv_obj;
 | |
| +		obj->variant.file_variant.stored_size = cp->size_or_equiv_obj;
 | |
| +	} else if (obj->variant_type == YAFFS_OBJECT_TYPE_HARDLINK) {
 | |
| +		obj->variant.hardlink_variant.equiv_id = cp->size_or_equiv_obj;
 | |
| +	}
 | |
| +	if (obj->hdr_chunk > 0)
 | |
| +		obj->lazy_loaded = 1;
 | |
| +	return 1;
 | |
| +}
 | |
| +
 | |
| +static void yaffs2_do_endian_tnode(struct yaffs_dev *dev, struct yaffs_tnode *tn)
 | |
| +{
 | |
| +	int i;
 | |
| +	u32 *as_u32 = (u32 *)tn;
 | |
| +	int tnode_size_u32 = dev->tnode_size / sizeof(u32);
 | |
| +
 | |
| +	if (!dev->swap_endian)
 | |
| +		return;
 | |
| +	/* Swap all the tnode data as u32s to fix endianisms. */
 | |
| +	for (i = 0; i<tnode_size_u32; i++)
 | |
| +		as_u32[i] = swap_u32(as_u32[i]);
 | |
| +}
 | |
| +
 | |
| +struct yaffs_tnode *yaffs2_do_endian_tnode_copy(struct yaffs_dev *dev,
 | |
| +					       struct yaffs_tnode *tn)
 | |
| +{
 | |
| +	if (!dev->swap_endian)
 | |
| +		return tn;
 | |
| +
 | |
| +	memcpy(dev->tn_swap_buffer, tn, dev->tnode_size);
 | |
| +	tn = dev->tn_swap_buffer;
 | |
| +
 | |
| +	yaffs2_do_endian_tnode(dev, tn);
 | |
| +
 | |
| +	return tn;
 | |
| +}
 | |
| +
 | |
| +static int yaffs2_checkpt_tnode_worker(struct yaffs_obj *in,
 | |
| +				       struct yaffs_tnode *tn, u32 level,
 | |
| +				       int chunk_offset)
 | |
| +{
 | |
| +	int i;
 | |
| +	struct yaffs_dev *dev = in->my_dev;
 | |
| +	int ok = 1;
 | |
| +	u32 base_offset;
 | |
| +
 | |
| +	if (!tn)
 | |
| +		return 1;
 | |
| +
 | |
| +	if (level > 0) {
 | |
| +		for (i = 0; i < YAFFS_NTNODES_INTERNAL && ok; i++) {
 | |
| +			if (!tn->internal[i])
 | |
| +				continue;
 | |
| +			ok = yaffs2_checkpt_tnode_worker(in,
 | |
| +				 tn->internal[i],
 | |
| +				 level - 1,
 | |
| +				 (chunk_offset <<
 | |
| +				  YAFFS_TNODES_INTERNAL_BITS) + i);
 | |
| +		}
 | |
| +		return ok;
 | |
| +	}
 | |
| +
 | |
| +	/* Level 0 tnode */
 | |
| +	base_offset = chunk_offset << YAFFS_TNODES_LEVEL0_BITS;
 | |
| +	yaffs_do_endian_u32(dev, &base_offset);
 | |
| +
 | |
| +	ok = (yaffs2_checkpt_wr(dev, &base_offset, sizeof(base_offset)) ==
 | |
| +			sizeof(base_offset));
 | |
| +	if (ok) {
 | |
| +		/*
 | |
| +		 * NB Can't do an in-place endian swizzle since that would
 | |
| +		 * damage current tnode data.
 | |
| +		 * If a tnode endian conversion is required we do a copy.
 | |
| +		 */
 | |
| +		tn = yaffs2_do_endian_tnode_copy(dev, tn);
 | |
| +		ok = (yaffs2_checkpt_wr(dev, tn, dev->tnode_size) ==
 | |
| +			(int)dev->tnode_size);
 | |
| +	}
 | |
| +	return ok;
 | |
| +}
 | |
| +
 | |
| +static int yaffs2_wr_checkpt_tnodes(struct yaffs_obj *obj)
 | |
| +{
 | |
| +	u32 end_marker = ~0;
 | |
| +	int ok = 1;
 | |
| +
 | |
| +	if (obj->variant_type != YAFFS_OBJECT_TYPE_FILE)
 | |
| +		return ok;
 | |
| +
 | |
| +	ok = yaffs2_checkpt_tnode_worker(obj,
 | |
| +					 obj->variant.file_variant.top,
 | |
| +					 obj->variant.file_variant.
 | |
| +					 top_level, 0);
 | |
| +	if (ok)
 | |
| +		ok = (yaffs2_checkpt_wr(obj->my_dev, &end_marker,
 | |
| +				sizeof(end_marker)) == sizeof(end_marker));
 | |
| +
 | |
| +	return ok ? 1 : 0;
 | |
| +}
 | |
| +
 | |
| +static int yaffs2_rd_checkpt_tnodes(struct yaffs_obj *obj)
 | |
| +{
 | |
| +	u32 base_chunk;
 | |
| +	int ok = 1;
 | |
| +	struct yaffs_dev *dev = obj->my_dev;
 | |
| +	struct yaffs_file_var *file_stuct_ptr = &obj->variant.file_variant;
 | |
| +	struct yaffs_tnode *tn;
 | |
| +	int nread = 0;
 | |
| +
 | |
| +	ok = (yaffs2_checkpt_rd(dev, &base_chunk, sizeof(base_chunk)) ==
 | |
| +	      sizeof(base_chunk));
 | |
| +
 | |
| +	yaffs_do_endian_u32(dev, &base_chunk);
 | |
| +
 | |
| +	while (ok && (~base_chunk)) {
 | |
| +		nread++;
 | |
| +		/* Read level 0 tnode */
 | |
| +
 | |
| +		tn = yaffs_get_tnode(dev);
 | |
| +		if (tn) {
 | |
| +			ok = (yaffs2_checkpt_rd(dev, tn, dev->tnode_size) ==
 | |
| +				(int)dev->tnode_size);
 | |
| +			yaffs2_do_endian_tnode(dev, tn);
 | |
| +		}
 | |
| +		else
 | |
| +			ok = 0;
 | |
| +
 | |
| +		if (tn && ok)
 | |
| +			ok = yaffs_add_find_tnode_0(dev,
 | |
| +						    file_stuct_ptr,
 | |
| +						    base_chunk, tn) ? 1 : 0;
 | |
| +
 | |
| +		if (ok) {
 | |
| +			ok = (yaffs2_checkpt_rd
 | |
| +			      (dev, &base_chunk,
 | |
| +			       sizeof(base_chunk)) == sizeof(base_chunk));
 | |
| +			yaffs_do_endian_u32(dev, &base_chunk);
 | |
| +		}
 | |
| +
 | |
| +	}
 | |
| +
 | |
| +	yaffs_trace(YAFFS_TRACE_CHECKPOINT,
 | |
| +		"Checkpoint read tnodes %d records, last %d. ok %d",
 | |
| +		nread, base_chunk, ok);
 | |
| +
 | |
| +	return ok ? 1 : 0;
 | |
| +}
 | |
| +
 | |
| +
 | |
| +static void yaffs2_do_endian_checkpt_obj(struct yaffs_dev *dev,
 | |
| +					 struct yaffs_checkpt_obj *cp)
 | |
| +{
 | |
| +	if (!dev->swap_endian)
 | |
| +		return;
 | |
| +	cp->struct_type = swap_s32(cp->struct_type);
 | |
| +	cp->obj_id = swap_u32(cp->obj_id);
 | |
| +	cp->parent_id = swap_u32(cp->parent_id);
 | |
| +	cp->hdr_chunk = swap_s32(cp->hdr_chunk);
 | |
| +	cp->bit_field = swap_u32(cp->bit_field);
 | |
| +	cp->n_data_chunks = swap_s32(cp->n_data_chunks);
 | |
| +	cp->size_or_equiv_obj = swap_loff_t(cp->size_or_equiv_obj);
 | |
| +}
 | |
| +
 | |
| +static int yaffs2_wr_checkpt_objs(struct yaffs_dev *dev)
 | |
| +{
 | |
| +	struct yaffs_obj *obj;
 | |
| +	struct yaffs_checkpt_obj cp;
 | |
| +	int i;
 | |
| +	int ok = 1;
 | |
| +	struct list_head *lh;
 | |
| +	u32 cp_variant_type;
 | |
| +
 | |
| +	/* Iterate through the objects in each hash entry,
 | |
| +	 * dumping them to the checkpointing stream.
 | |
| +	 */
 | |
| +
 | |
| +	(void) cp_variant_type;
 | |
| +
 | |
| +	for (i = 0; ok && i < YAFFS_NOBJECT_BUCKETS; i++) {
 | |
| +		list_for_each(lh, &dev->obj_bucket[i].list) {
 | |
| +			obj = list_entry(lh, struct yaffs_obj, hash_link);
 | |
| +			if (!obj->defered_free) {
 | |
| +				yaffs2_obj_checkpt_obj(&cp, obj);
 | |
| +				cp.struct_type = sizeof(cp);
 | |
| +				cp_variant_type = yaffs2_checkpt_obj_bit_get(
 | |
| +						&cp, CHECKPOINT_VARIANT_BITS);
 | |
| +				yaffs_trace(YAFFS_TRACE_CHECKPOINT,
 | |
| +					"Checkpoint write object %d parent %d type %d chunk %d obj addr %p",
 | |
| +					cp.obj_id, cp.parent_id,
 | |
| +					cp_variant_type, cp.hdr_chunk, obj);
 | |
| +
 | |
| +				yaffs2_do_endian_checkpt_obj (dev, &cp);
 | |
| +				ok = (yaffs2_checkpt_wr(dev, &cp,
 | |
| +						sizeof(cp)) == sizeof(cp));
 | |
| +
 | |
| +				if (ok &&
 | |
| +					obj->variant_type ==
 | |
| +					YAFFS_OBJECT_TYPE_FILE)
 | |
| +					ok = yaffs2_wr_checkpt_tnodes(obj);
 | |
| +			}
 | |
| +		}
 | |
| +	}
 | |
| +
 | |
| +	/* Dump end of list */
 | |
| +	memset(&cp, 0xff, sizeof(struct yaffs_checkpt_obj));
 | |
| +	cp.struct_type = sizeof(cp);
 | |
| +	yaffs2_do_endian_checkpt_obj (dev, &cp);
 | |
| +
 | |
| +	if (ok)
 | |
| +		ok = (yaffs2_checkpt_wr(dev, &cp, sizeof(cp)) == sizeof(cp));
 | |
| +
 | |
| +	return ok ? 1 : 0;
 | |
| +}
 | |
| +
 | |
| +static int yaffs2_rd_checkpt_objs(struct yaffs_dev *dev)
 | |
| +{
 | |
| +	struct yaffs_obj *obj;
 | |
| +	struct yaffs_checkpt_obj cp;
 | |
| +	int ok = 1;
 | |
| +	int done = 0;
 | |
| +	u32 cp_variant_type;
 | |
| +	LIST_HEAD(hard_list);
 | |
| +
 | |
| +
 | |
| +	while (ok && !done) {
 | |
| +		ok = (yaffs2_checkpt_rd(dev, &cp, sizeof(cp)) == sizeof(cp));
 | |
| +		yaffs2_do_endian_checkpt_obj (dev, &cp);
 | |
| +
 | |
| +		if (cp.struct_type != sizeof(cp)) {
 | |
| +			yaffs_trace(YAFFS_TRACE_CHECKPOINT,
 | |
| +				"struct size %d instead of %d ok %d",
 | |
| +				cp.struct_type, (int)sizeof(cp), ok);
 | |
| +			ok = 0;
 | |
| +		}
 | |
| +
 | |
| +		cp_variant_type = yaffs2_checkpt_obj_bit_get(
 | |
| +						&cp, CHECKPOINT_VARIANT_BITS);
 | |
| +		yaffs_trace(YAFFS_TRACE_CHECKPOINT,
 | |
| +			"Checkpoint read object %d parent %d type %d chunk %d ",
 | |
| +			cp.obj_id, cp.parent_id, cp_variant_type,
 | |
| +			cp.hdr_chunk);
 | |
| +
 | |
| +		if (ok && cp.obj_id == (u32)(~0)) {
 | |
| +			done = 1;
 | |
| +		} else if (ok) {
 | |
| +			obj =
 | |
| +			    yaffs_find_or_create_by_number(dev, cp.obj_id,
 | |
| +							   cp_variant_type);
 | |
| +			if (obj) {
 | |
| +				ok = yaffs2_checkpt_obj_to_obj(obj, &cp);
 | |
| +				if (!ok)
 | |
| +					break;
 | |
| +				if (obj->variant_type ==
 | |
| +					YAFFS_OBJECT_TYPE_FILE) {
 | |
| +					ok = yaffs2_rd_checkpt_tnodes(obj);
 | |
| +				} else if (obj->variant_type ==
 | |
| +					YAFFS_OBJECT_TYPE_HARDLINK) {
 | |
| +					list_add(&obj->hard_links, &hard_list);
 | |
| +				}
 | |
| +			} else {
 | |
| +				ok = 0;
 | |
| +			}
 | |
| +		}
 | |
| +	}
 | |
| +
 | |
| +	if (ok)
 | |
| +		yaffs_link_fixup(dev, &hard_list);
 | |
| +
 | |
| +	return ok ? 1 : 0;
 | |
| +}
 | |
| +
 | |
| +static int yaffs2_wr_checkpt_sum(struct yaffs_dev *dev)
 | |
| +{
 | |
| +	u32 checkpt_sum;
 | |
| +	int ok;
 | |
| +
 | |
| +	yaffs2_get_checkpt_sum(dev, &checkpt_sum);
 | |
| +
 | |
| +	yaffs_do_endian_u32(dev, &checkpt_sum);
 | |
| +
 | |
| +	ok = (yaffs2_checkpt_wr(dev, &checkpt_sum, sizeof(checkpt_sum)) ==
 | |
| +		sizeof(checkpt_sum));
 | |
| +
 | |
| +	if (!ok)
 | |
| +		return 0;
 | |
| +
 | |
| +	return 1;
 | |
| +}
 | |
| +
 | |
| +static int yaffs2_rd_checkpt_sum(struct yaffs_dev *dev)
 | |
| +{
 | |
| +	u32 checkpt_sum0;
 | |
| +	u32 checkpt_sum1;
 | |
| +	int ok;
 | |
| +
 | |
| +	yaffs2_get_checkpt_sum(dev, &checkpt_sum0);
 | |
| +
 | |
| +	ok = (yaffs2_checkpt_rd(dev, &checkpt_sum1, sizeof(checkpt_sum1)) ==
 | |
| +		sizeof(checkpt_sum1));
 | |
| +
 | |
| +	if (!ok)
 | |
| +		return 0;
 | |
| +	yaffs_do_endian_u32(dev, &checkpt_sum1);
 | |
| +
 | |
| +	if (checkpt_sum0 != checkpt_sum1)
 | |
| +		return 0;
 | |
| +
 | |
| +	return 1;
 | |
| +}
 | |
| +
 | |
| +static int yaffs2_wr_checkpt_data(struct yaffs_dev *dev)
 | |
| +{
 | |
| +	int ok = 1;
 | |
| +
 | |
| +	if (!yaffs2_checkpt_required(dev)) {
 | |
| +		yaffs_trace(YAFFS_TRACE_CHECKPOINT,
 | |
| +			"skipping checkpoint write");
 | |
| +		ok = 0;
 | |
| +	}
 | |
| +
 | |
| +	if (ok)
 | |
| +		ok = yaffs2_checkpt_open(dev, 1);
 | |
| +
 | |
| +	if (ok) {
 | |
| +		yaffs_trace(YAFFS_TRACE_CHECKPOINT,
 | |
| +			"write checkpoint validity");
 | |
| +		ok = yaffs2_wr_checkpt_validity_marker(dev, 1);
 | |
| +	}
 | |
| +	if (ok) {
 | |
| +		yaffs_trace(YAFFS_TRACE_CHECKPOINT,
 | |
| +			"write checkpoint device");
 | |
| +		ok = yaffs2_wr_checkpt_dev(dev);
 | |
| +	}
 | |
| +	if (ok) {
 | |
| +		yaffs_trace(YAFFS_TRACE_CHECKPOINT,
 | |
| +			"write checkpoint objects");
 | |
| +		ok = yaffs2_wr_checkpt_objs(dev);
 | |
| +	}
 | |
| +	if (ok) {
 | |
| +		yaffs_trace(YAFFS_TRACE_CHECKPOINT,
 | |
| +			"write checkpoint validity");
 | |
| +		ok = yaffs2_wr_checkpt_validity_marker(dev, 0);
 | |
| +	}
 | |
| +
 | |
| +	if (ok)
 | |
| +		ok = yaffs2_wr_checkpt_sum(dev);
 | |
| +
 | |
| +	if (!yaffs_checkpt_close(dev))
 | |
| +		ok = 0;
 | |
| +
 | |
| +	if (ok)
 | |
| +		dev->is_checkpointed = 1;
 | |
| +	else
 | |
| +		dev->is_checkpointed = 0;
 | |
| +
 | |
| +	return dev->is_checkpointed;
 | |
| +}
 | |
| +
 | |
| +static int yaffs2_rd_checkpt_data(struct yaffs_dev *dev)
 | |
| +{
 | |
| +	int ok = 1;
 | |
| +
 | |
| +	if (!dev->param.is_yaffs2)
 | |
| +		ok = 0;
 | |
| +
 | |
| +	if (ok && dev->param.skip_checkpt_rd) {
 | |
| +		yaffs_trace(YAFFS_TRACE_CHECKPOINT,
 | |
| +			"skipping checkpoint read");
 | |
| +		ok = 0;
 | |
| +	}
 | |
| +
 | |
| +	if (ok)
 | |
| +		ok = yaffs2_checkpt_open(dev, 0); /* open for read */
 | |
| +
 | |
| +	if (ok) {
 | |
| +		yaffs_trace(YAFFS_TRACE_CHECKPOINT,
 | |
| +			"read checkpoint validity");
 | |
| +		ok = yaffs2_rd_checkpt_validity_marker(dev, 1);
 | |
| +	}
 | |
| +	if (ok) {
 | |
| +		yaffs_trace(YAFFS_TRACE_CHECKPOINT,
 | |
| +			"read checkpoint device");
 | |
| +		ok = yaffs2_rd_checkpt_dev(dev);
 | |
| +	}
 | |
| +	if (ok) {
 | |
| +		yaffs_trace(YAFFS_TRACE_CHECKPOINT,
 | |
| +			"read checkpoint objects");
 | |
| +		ok = yaffs2_rd_checkpt_objs(dev);
 | |
| +	}
 | |
| +	if (ok) {
 | |
| +		yaffs_trace(YAFFS_TRACE_CHECKPOINT,
 | |
| +			"read checkpoint validity");
 | |
| +		ok = yaffs2_rd_checkpt_validity_marker(dev, 0);
 | |
| +	}
 | |
| +
 | |
| +	if (ok) {
 | |
| +		ok = yaffs2_rd_checkpt_sum(dev);
 | |
| +		yaffs_trace(YAFFS_TRACE_CHECKPOINT,
 | |
| +			"read checkpoint checksum %d", ok);
 | |
| +	}
 | |
| +
 | |
| +	if (!yaffs_checkpt_close(dev))
 | |
| +		ok = 0;
 | |
| +
 | |
| +	if (ok)
 | |
| +		dev->is_checkpointed = 1;
 | |
| +	else
 | |
| +		dev->is_checkpointed = 0;
 | |
| +
 | |
| +	return ok ? 1 : 0;
 | |
| +}
 | |
| +
 | |
| +void yaffs2_checkpt_invalidate(struct yaffs_dev *dev)
 | |
| +{
 | |
| +	if (dev->is_checkpointed || dev->blocks_in_checkpt > 0) {
 | |
| +		dev->is_checkpointed = 0;
 | |
| +		yaffs2_checkpt_invalidate_stream(dev);
 | |
| +	}
 | |
| +	if (dev->param.sb_dirty_fn)
 | |
| +		dev->param.sb_dirty_fn(dev);
 | |
| +}
 | |
| +
 | |
| +int yaffs_checkpoint_save(struct yaffs_dev *dev)
 | |
| +{
 | |
| +	yaffs_trace(YAFFS_TRACE_CHECKPOINT,
 | |
| +		"save entry: is_checkpointed %d",
 | |
| +		dev->is_checkpointed);
 | |
| +
 | |
| +	yaffs_verify_objects(dev);
 | |
| +	yaffs_verify_blocks(dev);
 | |
| +	yaffs_verify_free_chunks(dev);
 | |
| +
 | |
| +	if (!dev->is_checkpointed) {
 | |
| +		yaffs2_checkpt_invalidate(dev);
 | |
| +		yaffs2_wr_checkpt_data(dev);
 | |
| +	}
 | |
| +
 | |
| +	yaffs_trace(YAFFS_TRACE_CHECKPOINT | YAFFS_TRACE_MOUNT,
 | |
| +		"save exit: is_checkpointed %d",
 | |
| +		dev->is_checkpointed);
 | |
| +
 | |
| +	return dev->is_checkpointed;
 | |
| +}
 | |
| +
 | |
| +int yaffs2_checkpt_restore(struct yaffs_dev *dev)
 | |
| +{
 | |
| +	int retval;
 | |
| +
 | |
| +	yaffs_trace(YAFFS_TRACE_CHECKPOINT,
 | |
| +		"restore entry: is_checkpointed %d",
 | |
| +		dev->is_checkpointed);
 | |
| +
 | |
| +	retval = yaffs2_rd_checkpt_data(dev);
 | |
| +
 | |
| +	if (dev->is_checkpointed) {
 | |
| +		yaffs_verify_objects(dev);
 | |
| +		yaffs_verify_blocks(dev);
 | |
| +		yaffs_verify_free_chunks(dev);
 | |
| +	}
 | |
| +
 | |
| +	yaffs_trace(YAFFS_TRACE_CHECKPOINT,
 | |
| +		"restore exit: is_checkpointed %d",
 | |
| +		dev->is_checkpointed);
 | |
| +
 | |
| +	return retval;
 | |
| +}
 | |
| +
 | |
| +/* End of checkpointing */
 | |
| +
 | |
| +/* Hole handling logic for truncate past end of file */
 | |
| +
 | |
| +int yaffs2_handle_hole(struct yaffs_obj *obj, loff_t new_size)
 | |
| +{
 | |
| +	/* if new_size > old_file_size.
 | |
| +	 * We're going to be writing a hole.
 | |
| +	 * If the hole is small then write zeros otherwise write a start
 | |
| +	 * of hole marker.
 | |
| +	 */
 | |
| +	loff_t old_file_size;
 | |
| +	loff_t increase;
 | |
| +	int small_hole;
 | |
| +	int result = YAFFS_OK;
 | |
| +	struct yaffs_dev *dev = NULL;
 | |
| +	u8 *local_buffer = NULL;
 | |
| +	int small_increase_ok = 0;
 | |
| +
 | |
| +	if (!obj)
 | |
| +		return YAFFS_FAIL;
 | |
| +
 | |
| +	if (obj->variant_type != YAFFS_OBJECT_TYPE_FILE)
 | |
| +		return YAFFS_FAIL;
 | |
| +
 | |
| +	dev = obj->my_dev;
 | |
| +
 | |
| +	/* Bail out if not yaffs2 mode */
 | |
| +	if (!dev->param.is_yaffs2)
 | |
| +		return YAFFS_OK;
 | |
| +
 | |
| +	old_file_size = obj->variant.file_variant.file_size;
 | |
| +
 | |
| +	if (new_size <= old_file_size)
 | |
| +		return YAFFS_OK;
 | |
| +
 | |
| +	increase = new_size - old_file_size;
 | |
| +
 | |
| +	if (increase < YAFFS_SMALL_HOLE_THRESHOLD * dev->data_bytes_per_chunk &&
 | |
| +	    yaffs_check_alloc_available(dev, YAFFS_SMALL_HOLE_THRESHOLD + 1))
 | |
| +		small_hole = 1;
 | |
| +	else
 | |
| +		small_hole = 0;
 | |
| +
 | |
| +	if (small_hole)
 | |
| +		local_buffer = yaffs_get_temp_buffer(dev);
 | |
| +
 | |
| +	if (local_buffer) {
 | |
| +		/* fill hole with zero bytes */
 | |
| +		loff_t pos = old_file_size;
 | |
| +		int this_write;
 | |
| +		int written;
 | |
| +		memset(local_buffer, 0, dev->data_bytes_per_chunk);
 | |
| +		small_increase_ok = 1;
 | |
| +
 | |
| +		while (increase > 0 && small_increase_ok) {
 | |
| +			this_write = increase;
 | |
| +			if (this_write > (int)dev->data_bytes_per_chunk)
 | |
| +				this_write = dev->data_bytes_per_chunk;
 | |
| +			written =
 | |
| +			    yaffs_do_file_wr(obj, local_buffer, pos, this_write,
 | |
| +					     0);
 | |
| +			if (written == this_write) {
 | |
| +				pos += this_write;
 | |
| +				increase -= this_write;
 | |
| +			} else {
 | |
| +				small_increase_ok = 0;
 | |
| +			}
 | |
| +		}
 | |
| +
 | |
| +		yaffs_release_temp_buffer(dev, local_buffer);
 | |
| +
 | |
| +		/* If out of space then reverse any chunks we've added */
 | |
| +		if (!small_increase_ok)
 | |
| +			yaffs_resize_file_down(obj, old_file_size);
 | |
| +	}
 | |
| +
 | |
| +	if (!small_increase_ok &&
 | |
| +	    obj->parent &&
 | |
| +	    obj->parent->obj_id != YAFFS_OBJECTID_UNLINKED &&
 | |
| +	    obj->parent->obj_id != YAFFS_OBJECTID_DELETED) {
 | |
| +		/* Write a hole start header with the old file size */
 | |
| +		yaffs_update_oh(obj, NULL, 0, 1, 0, NULL);
 | |
| +	}
 | |
| +
 | |
| +	return result;
 | |
| +}
 | |
| +
 | |
| +/* Yaffs2 scanning */
 | |
| +
 | |
| +struct yaffs_block_index {
 | |
| +	int seq;
 | |
| +	int block;
 | |
| +};
 | |
| +
 | |
| +static int yaffs2_ybicmp(const void *a, const void *b)
 | |
| +{
 | |
| +	int aseq = ((struct yaffs_block_index *)a)->seq;
 | |
| +	int bseq = ((struct yaffs_block_index *)b)->seq;
 | |
| +	int ablock = ((struct yaffs_block_index *)a)->block;
 | |
| +	int bblock = ((struct yaffs_block_index *)b)->block;
 | |
| +
 | |
| +	if (aseq == bseq)
 | |
| +		return ablock - bblock;
 | |
| +
 | |
| +	return aseq - bseq;
 | |
| +}
 | |
| +
 | |
| +static inline int yaffs2_scan_chunk(struct yaffs_dev *dev,
 | |
| +		struct yaffs_block_info *bi,
 | |
| +		int blk, int chunk_in_block,
 | |
| +		int *found_chunks,
 | |
| +		u8 *chunk_data,
 | |
| +		struct list_head *hard_list,
 | |
| +		int summary_available)
 | |
| +{
 | |
| +	struct yaffs_obj_hdr *oh;
 | |
| +	struct yaffs_obj *in;
 | |
| +	struct yaffs_obj *parent;
 | |
| +	int equiv_id;
 | |
| +	loff_t file_size;
 | |
| +	int is_shrink;
 | |
| +	int is_unlinked;
 | |
| +	struct yaffs_ext_tags tags;
 | |
| +	int result;
 | |
| +	int alloc_failed = 0;
 | |
| +	int chunk = blk * dev->param.chunks_per_block + chunk_in_block;
 | |
| +	struct yaffs_file_var *file_var;
 | |
| +	struct yaffs_hardlink_var *hl_var;
 | |
| +	struct yaffs_symlink_var *sl_var;
 | |
| +
 | |
| +	if (summary_available) {
 | |
| +		result = yaffs_summary_fetch(dev, &tags, chunk_in_block);
 | |
| +		tags.seq_number = bi->seq_number;
 | |
| +	}
 | |
| +
 | |
| +	if (!summary_available || tags.obj_id == 0) {
 | |
| +		result = yaffs_rd_chunk_tags_nand(dev, chunk, NULL, &tags);
 | |
| +		dev->tags_used++;
 | |
| +	} else {
 | |
| +		dev->summary_used++;
 | |
| +	}
 | |
| +
 | |
| +	if (result == YAFFS_FAIL)
 | |
| +		yaffs_trace(YAFFS_TRACE_SCAN,
 | |
| +				"Could not get tags for chunk %d\n", chunk);
 | |
| +	/* Let's have a good look at this chunk... */
 | |
| +
 | |
| +	if (!tags.chunk_used) {
 | |
| +		/* An unassigned chunk in the block.
 | |
| +		 * If there are used chunks after this one, then
 | |
| +		 * it is a chunk that was skipped due to failing
 | |
| +		 * the erased check. Just skip it so that it can
 | |
| +		 * be deleted.
 | |
| +		 * But, more typically, We get here when this is
 | |
| +		 * an unallocated chunk and his means that
 | |
| +		 * either the block is empty or this is the one
 | |
| +		 * being allocated from
 | |
| +		 */
 | |
| +
 | |
| +		if (*found_chunks) {
 | |
| +			/* This is a chunk that was skipped due
 | |
| +			 * to failing the erased check */
 | |
| +		} else if (chunk_in_block == 0) {
 | |
| +			/* We're looking at the first chunk in
 | |
| +			 * the block so the block is unused */
 | |
| +			bi->block_state = YAFFS_BLOCK_STATE_EMPTY;
 | |
| +			dev->n_erased_blocks++;
 | |
| +		} else {
 | |
| +			if (bi->block_state == YAFFS_BLOCK_STATE_NEEDS_SCAN ||
 | |
| +			    bi->block_state == YAFFS_BLOCK_STATE_ALLOCATING) {
 | |
| +				if (dev->seq_number == bi->seq_number) {
 | |
| +					/* Allocating from this block*/
 | |
| +					yaffs_trace(YAFFS_TRACE_SCAN,
 | |
| +					    " Allocating from %d %d",
 | |
| +					    blk, chunk_in_block);
 | |
| +
 | |
| +					bi->block_state =
 | |
| +						YAFFS_BLOCK_STATE_ALLOCATING;
 | |
| +					dev->alloc_block = blk;
 | |
| +					dev->alloc_page = chunk_in_block;
 | |
| +					dev->alloc_block_finder = blk;
 | |
| +				} else {
 | |
| +					/* This is a partially written block
 | |
| +					 * that is not the current
 | |
| +					 * allocation block.
 | |
| +					 */
 | |
| +					yaffs_trace(YAFFS_TRACE_SCAN,
 | |
| +						"Partially written block %d detected. gc will fix this.",
 | |
| +						blk);
 | |
| +				}
 | |
| +			}
 | |
| +		}
 | |
| +
 | |
| +		dev->n_free_chunks++;
 | |
| +
 | |
| +	} else if (tags.ecc_result ==
 | |
| +		YAFFS_ECC_RESULT_UNFIXED) {
 | |
| +		yaffs_trace(YAFFS_TRACE_SCAN,
 | |
| +			" Unfixed ECC in chunk(%d:%d), chunk ignored",
 | |
| +			blk, chunk_in_block);
 | |
| +			dev->n_free_chunks++;
 | |
| +	} else if (tags.obj_id > YAFFS_MAX_OBJECT_ID ||
 | |
| +		   tags.chunk_id > YAFFS_MAX_CHUNK_ID ||
 | |
| +		   tags.obj_id == YAFFS_OBJECTID_SUMMARY ||
 | |
| +		   (tags.chunk_id > 0 &&
 | |
| +		     tags.n_bytes > dev->data_bytes_per_chunk) ||
 | |
| +		   tags.seq_number != bi->seq_number) {
 | |
| +		yaffs_trace(YAFFS_TRACE_SCAN,
 | |
| +			"Chunk (%d:%d) with bad tags:obj = %d, chunk_id = %d, n_bytes = %d, ignored",
 | |
| +			blk, chunk_in_block, tags.obj_id,
 | |
| +			tags.chunk_id, tags.n_bytes);
 | |
| +		dev->n_free_chunks++;
 | |
| +	} else if (tags.chunk_id > 0) {
 | |
| +		/* chunk_id > 0 so it is a data chunk... */
 | |
| +		loff_t endpos;
 | |
| +		loff_t chunk_base = (tags.chunk_id - 1) *
 | |
| +					dev->data_bytes_per_chunk;
 | |
| +
 | |
| +		*found_chunks = 1;
 | |
| +
 | |
| +		yaffs_set_chunk_bit(dev, blk, chunk_in_block);
 | |
| +		bi->pages_in_use++;
 | |
| +
 | |
| +		in = yaffs_find_or_create_by_number(dev,
 | |
| +					tags.obj_id,
 | |
| +					YAFFS_OBJECT_TYPE_FILE);
 | |
| +		if (!in)
 | |
| +			/* Out of memory */
 | |
| +			alloc_failed = 1;
 | |
| +
 | |
| +		if (in &&
 | |
| +		    in->variant_type == YAFFS_OBJECT_TYPE_FILE &&
 | |
| +		    chunk_base < in->variant.file_variant.shrink_size) {
 | |
| +			/* This has not been invalidated by
 | |
| +			 * a resize */
 | |
| +			if (!yaffs_put_chunk_in_file(in, tags.chunk_id,
 | |
| +								chunk, -1))
 | |
| +				alloc_failed = 1;
 | |
| +
 | |
| +			/* File size is calculated by looking at
 | |
| +			 * the data chunks if we have not
 | |
| +			 * seen an object header yet.
 | |
| +			 * Stop this practice once we find an
 | |
| +			 * object header.
 | |
| +			 */
 | |
| +			endpos = chunk_base + tags.n_bytes;
 | |
| +
 | |
| +			if (!in->valid &&
 | |
| +			    in->variant.file_variant.stored_size < endpos) {
 | |
| +				in->variant.file_variant.
 | |
| +				    stored_size = endpos;
 | |
| +				in->variant.file_variant.
 | |
| +				    file_size = endpos;
 | |
| +			}
 | |
| +		} else if (in) {
 | |
| +			/* This chunk has been invalidated by a
 | |
| +			 * resize, or a past file deletion
 | |
| +			 * so delete the chunk*/
 | |
| +			yaffs_chunk_del(dev, chunk, 1, __LINE__);
 | |
| +		}
 | |
| +	} else {
 | |
| +		/* chunk_id == 0, so it is an ObjectHeader.
 | |
| +		 * Thus, we read in the object header and make
 | |
| +		 * the object
 | |
| +		 */
 | |
| +		*found_chunks = 1;
 | |
| +
 | |
| +		yaffs_set_chunk_bit(dev, blk, chunk_in_block);
 | |
| +		bi->pages_in_use++;
 | |
| +
 | |
| +		oh = NULL;
 | |
| +		in = NULL;
 | |
| +
 | |
| +		if (tags.extra_available) {
 | |
| +			in = yaffs_find_or_create_by_number(dev,
 | |
| +					tags.obj_id,
 | |
| +					tags.extra_obj_type);
 | |
| +			if (!in)
 | |
| +				alloc_failed = 1;
 | |
| +		}
 | |
| +
 | |
| +		if (!in ||
 | |
| +		    (!in->valid && dev->param.disable_lazy_load) ||
 | |
| +		    tags.extra_shadows ||
 | |
| +		    (!in->valid && (tags.obj_id == YAFFS_OBJECTID_ROOT ||
 | |
| +				 tags.obj_id == YAFFS_OBJECTID_LOSTNFOUND))) {
 | |
| +
 | |
| +			/* If we don't have  valid info then we
 | |
| +			 * need to read the chunk
 | |
| +			 * TODO In future we can probably defer
 | |
| +			 * reading the chunk and living with
 | |
| +			 * invalid data until needed.
 | |
| +			 */
 | |
| +
 | |
| +			result = yaffs_rd_chunk_tags_nand(dev,
 | |
| +						  chunk,
 | |
| +						  chunk_data,
 | |
| +						  NULL);
 | |
| +
 | |
| +			oh = (struct yaffs_obj_hdr *)chunk_data;
 | |
| +
 | |
| +			yaffs_do_endian_oh(dev, oh);
 | |
| +
 | |
| +			if (dev->param.inband_tags) {
 | |
| +				/* Fix up the header if they got
 | |
| +				 * corrupted by inband tags */
 | |
| +				oh->shadows_obj =
 | |
| +				    oh->inband_shadowed_obj_id;
 | |
| +				oh->is_shrink =
 | |
| +				    oh->inband_is_shrink;
 | |
| +			}
 | |
| +
 | |
| +			if (!in) {
 | |
| +				in = yaffs_find_or_create_by_number(dev,
 | |
| +							tags.obj_id, oh->type);
 | |
| +				if (!in)
 | |
| +					alloc_failed = 1;
 | |
| +			}
 | |
| +		}
 | |
| +
 | |
| +		if (!in) {
 | |
| +			/* TODO Hoosterman we have a problem! */
 | |
| +			yaffs_trace(YAFFS_TRACE_ERROR,
 | |
| +				"yaffs tragedy: Could not make object for object  %d at chunk %d during scan",
 | |
| +				tags.obj_id, chunk);
 | |
| +			return YAFFS_FAIL;
 | |
| +		}
 | |
| +
 | |
| +		if (in->valid) {
 | |
| +			/* We have already filled this one.
 | |
| +			 * We have a duplicate that will be
 | |
| +			 * discarded, but we first have to suck
 | |
| +			 * out resize info if it is a file.
 | |
| +			 */
 | |
| +			if ((in->variant_type == YAFFS_OBJECT_TYPE_FILE) &&
 | |
| +				((oh && oh->type == YAFFS_OBJECT_TYPE_FILE) ||
 | |
| +				 (tags.extra_available &&
 | |
| +				  tags.extra_obj_type == YAFFS_OBJECT_TYPE_FILE)
 | |
| +				)) {
 | |
| +				loff_t this_size = (oh) ?
 | |
| +					yaffs_oh_to_size(dev, oh, 0) :
 | |
| +					tags.extra_file_size;
 | |
| +				u32 parent_obj_id = (oh) ?
 | |
| +					(u32)oh->parent_obj_id :
 | |
| +					tags.extra_parent_id;
 | |
| +
 | |
| +				is_shrink = (oh) ?
 | |
| +					oh->is_shrink :
 | |
| +					tags.extra_is_shrink;
 | |
| +
 | |
| +				/* If it is deleted (unlinked
 | |
| +				 * at start also means deleted)
 | |
| +				 * we treat the file size as
 | |
| +				 * being zeroed at this point.
 | |
| +				 */
 | |
| +				if (parent_obj_id == YAFFS_OBJECTID_DELETED ||
 | |
| +				    parent_obj_id == YAFFS_OBJECTID_UNLINKED) {
 | |
| +					this_size = 0;
 | |
| +					is_shrink = 1;
 | |
| +				}
 | |
| +
 | |
| +				if (is_shrink &&
 | |
| +				    in->variant.file_variant.shrink_size >
 | |
| +				    this_size)
 | |
| +					in->variant.file_variant.shrink_size =
 | |
| +					this_size;
 | |
| +
 | |
| +				if (is_shrink)
 | |
| +					bi->has_shrink_hdr = 1;
 | |
| +			}
 | |
| +			/* Use existing - destroy this one. */
 | |
| +			yaffs_chunk_del(dev, chunk, 1, __LINE__);
 | |
| +		}
 | |
| +
 | |
| +		if (!in->valid && in->variant_type !=
 | |
| +		    (oh ? oh->type : tags.extra_obj_type)) {
 | |
| +			yaffs_trace(YAFFS_TRACE_ERROR,
 | |
| +				"yaffs tragedy: Bad type, %d != %d, for object %d at chunk %d during scan",
 | |
| +				oh ? oh->type : tags.extra_obj_type,
 | |
| +				in->variant_type, tags.obj_id,
 | |
| +				chunk);
 | |
| +			in = yaffs_retype_obj(in, oh ? oh->type : tags.extra_obj_type);
 | |
| +		}
 | |
| +
 | |
| +		if (!in->valid &&
 | |
| +		    (tags.obj_id == YAFFS_OBJECTID_ROOT ||
 | |
| +		     tags.obj_id == YAFFS_OBJECTID_LOSTNFOUND)) {
 | |
| +			/* We only load some info, don't fiddle
 | |
| +			 * with directory structure */
 | |
| +			in->valid = 1;
 | |
| +
 | |
| +			if (oh) {
 | |
| +				in->yst_mode = oh->yst_mode;
 | |
| +				yaffs_load_attribs(in, oh);
 | |
| +				in->lazy_loaded = 0;
 | |
| +			} else {
 | |
| +				in->lazy_loaded = 1;
 | |
| +			}
 | |
| +			in->hdr_chunk = chunk;
 | |
| +
 | |
| +		} else if (!in->valid) {
 | |
| +			/* we need to load this info */
 | |
| +			in->valid = 1;
 | |
| +			in->hdr_chunk = chunk;
 | |
| +			if (oh) {
 | |
| +				in->variant_type = oh->type;
 | |
| +				in->yst_mode = oh->yst_mode;
 | |
| +				yaffs_load_attribs(in, oh);
 | |
| +
 | |
| +				if (oh->shadows_obj > 0)
 | |
| +					yaffs_handle_shadowed_obj(dev,
 | |
| +					     oh->shadows_obj, 1);
 | |
| +
 | |
| +				yaffs_set_obj_name_from_oh(in, oh);
 | |
| +				parent = yaffs_find_or_create_by_number(dev,
 | |
| +						oh->parent_obj_id,
 | |
| +						YAFFS_OBJECT_TYPE_DIRECTORY);
 | |
| +				file_size = yaffs_oh_to_size(dev, oh, 0);
 | |
| +				is_shrink = oh->is_shrink;
 | |
| +				equiv_id = oh->equiv_id;
 | |
| +			} else {
 | |
| +				in->variant_type = tags.extra_obj_type;
 | |
| +				parent = yaffs_find_or_create_by_number(dev,
 | |
| +						tags.extra_parent_id,
 | |
| +						YAFFS_OBJECT_TYPE_DIRECTORY);
 | |
| +				file_size = tags.extra_file_size;
 | |
| +				is_shrink = tags.extra_is_shrink;
 | |
| +				equiv_id = tags.extra_equiv_id;
 | |
| +				in->lazy_loaded = 1;
 | |
| +			}
 | |
| +			in->dirty = 0;
 | |
| +
 | |
| +			if (!parent)
 | |
| +				alloc_failed = 1;
 | |
| +
 | |
| +			/* directory stuff...
 | |
| +			 * hook up to parent
 | |
| +			 */
 | |
| +
 | |
| +			if (parent &&
 | |
| +			    parent->variant_type == YAFFS_OBJECT_TYPE_UNKNOWN) {
 | |
| +				/* Set up as a directory */
 | |
| +				parent->variant_type =
 | |
| +					YAFFS_OBJECT_TYPE_DIRECTORY;
 | |
| +				INIT_LIST_HEAD(&parent->
 | |
| +						variant.dir_variant.children);
 | |
| +			} else if (!parent ||
 | |
| +				   parent->variant_type !=
 | |
| +					YAFFS_OBJECT_TYPE_DIRECTORY) {
 | |
| +				/* Hoosterman, another problem....
 | |
| +				 * Trying to use a non-directory as a directory
 | |
| +				 */
 | |
| +
 | |
| +				yaffs_trace(YAFFS_TRACE_ERROR,
 | |
| +					"yaffs tragedy: attempting to use non-directory as a directory in scan. Put in lost+found."
 | |
| +					);
 | |
| +				parent = dev->lost_n_found;
 | |
| +			}
 | |
| +			yaffs_add_obj_to_dir(parent, in);
 | |
| +
 | |
| +			is_unlinked = (parent == dev->del_dir) ||
 | |
| +					(parent == dev->unlinked_dir);
 | |
| +
 | |
| +			if (is_shrink)
 | |
| +				/* Mark the block */
 | |
| +				bi->has_shrink_hdr = 1;
 | |
| +
 | |
| +			/* Note re hardlinks.
 | |
| +			 * Since we might scan a hardlink before its equivalent
 | |
| +			 * object is scanned we put them all in a list.
 | |
| +			 * After scanning is complete, we should have all the
 | |
| +			 * objects, so we run through this list and fix up all
 | |
| +			 * the chains.
 | |
| +			 */
 | |
| +
 | |
| +			switch (in->variant_type) {
 | |
| +			case YAFFS_OBJECT_TYPE_UNKNOWN:
 | |
| +				/* Todo got a problem */
 | |
| +				break;
 | |
| +			case YAFFS_OBJECT_TYPE_FILE:
 | |
| +				file_var = &in->variant.file_variant;
 | |
| +				if (file_var->stored_size < file_size) {
 | |
| +					/* This covers the case where the file
 | |
| +					 * size is greater than the data held.
 | |
| +					 * This will happen if the file is
 | |
| +					 * resized to be larger than its
 | |
| +					 * current data extents.
 | |
| +					 */
 | |
| +					file_var->file_size = file_size;
 | |
| +					file_var->stored_size = file_size;
 | |
| +				}
 | |
| +
 | |
| +				if (file_var->shrink_size > file_size)
 | |
| +					file_var->shrink_size = file_size;
 | |
| +
 | |
| +				break;
 | |
| +			case YAFFS_OBJECT_TYPE_HARDLINK:
 | |
| +				hl_var = &in->variant.hardlink_variant;
 | |
| +				if (!is_unlinked) {
 | |
| +					hl_var->equiv_id = equiv_id;
 | |
| +					list_add(&in->hard_links, hard_list);
 | |
| +				}
 | |
| +				break;
 | |
| +			case YAFFS_OBJECT_TYPE_DIRECTORY:
 | |
| +				/* Do nothing */
 | |
| +				break;
 | |
| +			case YAFFS_OBJECT_TYPE_SPECIAL:
 | |
| +				/* Do nothing */
 | |
| +				break;
 | |
| +			case YAFFS_OBJECT_TYPE_SYMLINK:
 | |
| +				sl_var = &in->variant.symlink_variant;
 | |
| +				if (oh) {
 | |
| +					sl_var->alias =
 | |
| +					    yaffs_clone_str(oh->alias);
 | |
| +					if (!sl_var->alias)
 | |
| +						alloc_failed = 1;
 | |
| +				}
 | |
| +				break;
 | |
| +			}
 | |
| +		}
 | |
| +	}
 | |
| +	return alloc_failed ? YAFFS_FAIL : YAFFS_OK;
 | |
| +}
 | |
| +
 | |
| +int yaffs2_scan_backwards(struct yaffs_dev *dev)
 | |
| +{
 | |
| +	u32 blk;
 | |
| +	int block_iter;
 | |
| +	int start_iter;
 | |
| +	int end_iter;
 | |
| +	int n_to_scan = 0;
 | |
| +	enum yaffs_block_state state;
 | |
| +	int c;
 | |
| +	LIST_HEAD(hard_list);
 | |
| +	struct yaffs_block_info *bi;
 | |
| +	u32 seq_number;
 | |
| +	int n_blocks = dev->internal_end_block - dev->internal_start_block + 1;
 | |
| +	u8 *chunk_data;
 | |
| +	int found_chunks;
 | |
| +	int alloc_failed = 0;
 | |
| +	struct yaffs_block_index *block_index = NULL;
 | |
| +	int alt_block_index = 0;
 | |
| +	int summary_available;
 | |
| +
 | |
| +	yaffs_trace(YAFFS_TRACE_SCAN,
 | |
| +		"yaffs2_scan_backwards starts  intstartblk %d intendblk %d...",
 | |
| +		dev->internal_start_block, dev->internal_end_block);
 | |
| +
 | |
| +	dev->seq_number = YAFFS_LOWEST_SEQUENCE_NUMBER;
 | |
| +
 | |
| +	block_index =
 | |
| +		kmalloc(n_blocks * sizeof(struct yaffs_block_index), GFP_NOFS);
 | |
| +
 | |
| +	if (!block_index) {
 | |
| +		block_index =
 | |
| +		    vmalloc(n_blocks * sizeof(struct yaffs_block_index));
 | |
| +		alt_block_index = 1;
 | |
| +	}
 | |
| +
 | |
| +	if (!block_index) {
 | |
| +		yaffs_trace(YAFFS_TRACE_SCAN,
 | |
| +			"yaffs2_scan_backwards() could not allocate block index!"
 | |
| +			);
 | |
| +		return YAFFS_FAIL;
 | |
| +	}
 | |
| +
 | |
| +	dev->blocks_in_checkpt = 0;
 | |
| +
 | |
| +	chunk_data = yaffs_get_temp_buffer(dev);
 | |
| +
 | |
| +	/* Scan all the blocks to determine their state */
 | |
| +	bi = dev->block_info;
 | |
| +	for (blk = dev->internal_start_block; blk <= dev->internal_end_block;
 | |
| +	     blk++) {
 | |
| +		yaffs_clear_chunk_bits(dev, blk);
 | |
| +		bi->pages_in_use = 0;
 | |
| +		bi->soft_del_pages = 0;
 | |
| +
 | |
| +		yaffs_query_init_block_state(dev, blk, &state, &seq_number);
 | |
| +
 | |
| +		bi->block_state = state;
 | |
| +		bi->seq_number = seq_number;
 | |
| +
 | |
| +		if (bi->seq_number == YAFFS_SEQUENCE_CHECKPOINT_DATA)
 | |
| +			bi->block_state = YAFFS_BLOCK_STATE_CHECKPOINT;
 | |
| +		if (bi->seq_number == YAFFS_SEQUENCE_BAD_BLOCK)
 | |
| +			bi->block_state = YAFFS_BLOCK_STATE_DEAD;
 | |
| +
 | |
| +		yaffs_trace(YAFFS_TRACE_SCAN_DEBUG,
 | |
| +			"Block scanning block %d state %d seq %d",
 | |
| +			blk, bi->block_state, seq_number);
 | |
| +
 | |
| +		if (bi->block_state == YAFFS_BLOCK_STATE_CHECKPOINT) {
 | |
| +			dev->blocks_in_checkpt++;
 | |
| +
 | |
| +		} else if (bi->block_state == YAFFS_BLOCK_STATE_DEAD) {
 | |
| +			yaffs_trace(YAFFS_TRACE_BAD_BLOCKS,
 | |
| +				"block %d is bad", blk);
 | |
| +		} else if (bi->block_state == YAFFS_BLOCK_STATE_EMPTY) {
 | |
| +			yaffs_trace(YAFFS_TRACE_SCAN_DEBUG, "Block empty ");
 | |
| +			dev->n_erased_blocks++;
 | |
| +			dev->n_free_chunks += dev->param.chunks_per_block;
 | |
| +		} else if (bi->block_state ==
 | |
| +				YAFFS_BLOCK_STATE_NEEDS_SCAN) {
 | |
| +			/* Determine the highest sequence number */
 | |
| +			if (seq_number >= YAFFS_LOWEST_SEQUENCE_NUMBER &&
 | |
| +			    seq_number < YAFFS_HIGHEST_SEQUENCE_NUMBER) {
 | |
| +				block_index[n_to_scan].seq = seq_number;
 | |
| +				block_index[n_to_scan].block = blk;
 | |
| +				n_to_scan++;
 | |
| +				if (seq_number >= dev->seq_number)
 | |
| +					dev->seq_number = seq_number;
 | |
| +			} else {
 | |
| +				/* TODO: Nasty sequence number! */
 | |
| +				yaffs_trace(YAFFS_TRACE_SCAN,
 | |
| +					"Block scanning block %d has bad sequence number %d",
 | |
| +					blk, seq_number);
 | |
| +			}
 | |
| +		}
 | |
| +		bi++;
 | |
| +	}
 | |
| +
 | |
| +	yaffs_trace(YAFFS_TRACE_ALWAYS, "%d blocks to be sorted...", n_to_scan);
 | |
| +
 | |
| +	cond_resched();
 | |
| +
 | |
| +	/* Sort the blocks by sequence number */
 | |
| +	sort(block_index, n_to_scan, sizeof(struct yaffs_block_index),
 | |
| +		   yaffs2_ybicmp, NULL);
 | |
| +
 | |
| +	cond_resched();
 | |
| +
 | |
| +	yaffs_trace(YAFFS_TRACE_SCAN, "...done");
 | |
| +
 | |
| +	/* Now scan the blocks looking at the data. */
 | |
| +	start_iter = 0;
 | |
| +	end_iter = n_to_scan - 1;
 | |
| +	yaffs_trace(YAFFS_TRACE_SCAN_DEBUG, "%d blocks to scan", n_to_scan);
 | |
| +
 | |
| +	/* For each block.... backwards */
 | |
| +	for (block_iter = end_iter;
 | |
| +	     !alloc_failed && block_iter >= start_iter;
 | |
| +	     block_iter--) {
 | |
| +		/* Cooperative multitasking! This loop can run for so
 | |
| +		   long that watchdog timers expire. */
 | |
| +		cond_resched();
 | |
| +
 | |
| +		/* get the block to scan in the correct order */
 | |
| +		blk = block_index[block_iter].block;
 | |
| +		bi = yaffs_get_block_info(dev, blk);
 | |
| +
 | |
| +		summary_available = yaffs_summary_read(dev, dev->sum_tags, blk);
 | |
| +
 | |
| +		/* For each chunk in each block that needs scanning.... */
 | |
| +		found_chunks = 0;
 | |
| +		if (summary_available)
 | |
| +			c = dev->chunks_per_summary - 1;
 | |
| +		else
 | |
| +			c = dev->param.chunks_per_block - 1;
 | |
| +
 | |
| +		for (/* c is already initialised */;
 | |
| +		     !alloc_failed && c >= 0 &&
 | |
| +		     (bi->block_state == YAFFS_BLOCK_STATE_NEEDS_SCAN ||
 | |
| +		      bi->block_state == YAFFS_BLOCK_STATE_ALLOCATING);
 | |
| +		      c--) {
 | |
| +			/* Scan backwards...
 | |
| +			 * Read the tags and decide what to do
 | |
| +			 */
 | |
| +			if (yaffs2_scan_chunk(dev, bi, blk, c,
 | |
| +					&found_chunks, chunk_data,
 | |
| +					&hard_list, summary_available) ==
 | |
| +					YAFFS_FAIL)
 | |
| +				alloc_failed = 1;
 | |
| +		}
 | |
| +
 | |
| +		if (bi->block_state == YAFFS_BLOCK_STATE_NEEDS_SCAN) {
 | |
| +			/* If we got this far while scanning, then the block
 | |
| +			 * is fully allocated. */
 | |
| +			bi->block_state = YAFFS_BLOCK_STATE_FULL;
 | |
| +		}
 | |
| +
 | |
| +		/* Now let's see if it was dirty */
 | |
| +		if (bi->pages_in_use == 0 &&
 | |
| +		    !bi->has_shrink_hdr &&
 | |
| +		    bi->block_state == YAFFS_BLOCK_STATE_FULL) {
 | |
| +			yaffs_block_became_dirty(dev, blk);
 | |
| +		}
 | |
| +	}
 | |
| +
 | |
| +	yaffs_skip_rest_of_block(dev);
 | |
| +
 | |
| +	if (alt_block_index)
 | |
| +		vfree(block_index);
 | |
| +	else
 | |
| +		kfree(block_index);
 | |
| +
 | |
| +	/* Ok, we've done all the scanning.
 | |
| +	 * Fix up the hard link chains.
 | |
| +	 * We have scanned all the objects, now it's time to add these
 | |
| +	 * hardlinks.
 | |
| +	 */
 | |
| +	yaffs_link_fixup(dev, &hard_list);
 | |
| +
 | |
| +	yaffs_release_temp_buffer(dev, chunk_data);
 | |
| +
 | |
| +	if (alloc_failed)
 | |
| +		return YAFFS_FAIL;
 | |
| +
 | |
| +	yaffs_trace(YAFFS_TRACE_SCAN, "yaffs2_scan_backwards ends");
 | |
| +
 | |
| +	return YAFFS_OK;
 | |
| +}
 |