mirror of https://github.com/OpenIPC/firmware.git
1718 lines
45 KiB
Diff
1718 lines
45 KiB
Diff
diff -drupN a/fs/yaffs2/yaffs_yaffs2.c b/fs/yaffs2/yaffs_yaffs2.c
|
|
--- a/fs/yaffs2/yaffs_yaffs2.c 1970-01-01 03:00:00.000000000 +0300
|
|
+++ b/fs/yaffs2/yaffs_yaffs2.c 2022-06-09 05:02:35.000000000 +0300
|
|
@@ -0,0 +1,1713 @@
|
|
+/*
|
|
+ * YAFFS: Yet Another Flash File System. A NAND-flash specific file system.
|
|
+ *
|
|
+ * Copyright (C) 2002-2018 Aleph One Ltd.
|
|
+ *
|
|
+ * Created by Charles Manning <charles@aleph1.co.uk>
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or modify
|
|
+ * it under the terms of the GNU General Public License version 2 as
|
|
+ * published by the Free Software Foundation.
|
|
+ */
|
|
+
|
|
+#include "yaffs_guts.h"
|
|
+#include "yaffs_trace.h"
|
|
+#include "yaffs_yaffs2.h"
|
|
+#include "yaffs_checkptrw.h"
|
|
+#include "yaffs_bitmap.h"
|
|
+#include "yaffs_nand.h"
|
|
+#include "yaffs_getblockinfo.h"
|
|
+#include "yaffs_verify.h"
|
|
+#include "yaffs_attribs.h"
|
|
+#include "yaffs_summary.h"
|
|
+#include "yaffs_endian.h"
|
|
+
|
|
+/*
|
|
+ * Checkpoints are really no benefit on very small partitions.
|
|
+ *
|
|
+ * To save space on small partitions don't bother with checkpoints unless
|
|
+ * the partition is at least this big.
|
|
+ */
|
|
+#define YAFFS_CHECKPOINT_MIN_BLOCKS 60
|
|
+#define YAFFS_SMALL_HOLE_THRESHOLD 4
|
|
+
|
|
+/*
|
|
+ * Oldest Dirty Sequence Number handling.
|
|
+ */
|
|
+
|
|
+/* yaffs_calc_oldest_dirty_seq()
|
|
+ * yaffs2_find_oldest_dirty_seq()
|
|
+ * Calculate the oldest dirty sequence number if we don't know it.
|
|
+ */
|
|
+void yaffs_calc_oldest_dirty_seq(struct yaffs_dev *dev)
|
|
+{
|
|
+ u32 i;
|
|
+ unsigned seq;
|
|
+ unsigned block_no = 0;
|
|
+ struct yaffs_block_info *b;
|
|
+
|
|
+ if (!dev->param.is_yaffs2)
|
|
+ return;
|
|
+
|
|
+ /* Find the oldest dirty sequence number. */
|
|
+ seq = dev->seq_number + 1;
|
|
+ b = dev->block_info;
|
|
+ for (i = dev->internal_start_block; i <= dev->internal_end_block; i++) {
|
|
+ if (b->block_state == YAFFS_BLOCK_STATE_FULL &&
|
|
+ (u32)(b->pages_in_use - b->soft_del_pages) <
|
|
+ dev->param.chunks_per_block &&
|
|
+ b->seq_number < seq) {
|
|
+ seq = b->seq_number;
|
|
+ block_no = i;
|
|
+ }
|
|
+ b++;
|
|
+ }
|
|
+
|
|
+ if (block_no) {
|
|
+ dev->oldest_dirty_seq = seq;
|
|
+ dev->oldest_dirty_block = block_no;
|
|
+ }
|
|
+}
|
|
+
|
|
+void yaffs2_find_oldest_dirty_seq(struct yaffs_dev *dev)
|
|
+{
|
|
+ if (!dev->param.is_yaffs2)
|
|
+ return;
|
|
+
|
|
+ if (!dev->oldest_dirty_seq)
|
|
+ yaffs_calc_oldest_dirty_seq(dev);
|
|
+}
|
|
+
|
|
+/*
|
|
+ * yaffs_clear_oldest_dirty_seq()
|
|
+ * Called when a block is erased or marked bad. (ie. when its seq_number
|
|
+ * becomes invalid). If the value matches the oldest then we clear
|
|
+ * dev->oldest_dirty_seq to force its recomputation.
|
|
+ */
|
|
+void yaffs2_clear_oldest_dirty_seq(struct yaffs_dev *dev,
|
|
+ struct yaffs_block_info *bi)
|
|
+{
|
|
+
|
|
+ if (!dev->param.is_yaffs2)
|
|
+ return;
|
|
+
|
|
+ if (!bi || bi->seq_number == dev->oldest_dirty_seq) {
|
|
+ dev->oldest_dirty_seq = 0;
|
|
+ dev->oldest_dirty_block = 0;
|
|
+ }
|
|
+}
|
|
+
|
|
+/*
|
|
+ * yaffs2_update_oldest_dirty_seq()
|
|
+ * Update the oldest dirty sequence number whenever we dirty a block.
|
|
+ * Only do this if the oldest_dirty_seq is actually being tracked.
|
|
+ */
|
|
+void yaffs2_update_oldest_dirty_seq(struct yaffs_dev *dev, unsigned block_no,
|
|
+ struct yaffs_block_info *bi)
|
|
+{
|
|
+ if (!dev->param.is_yaffs2)
|
|
+ return;
|
|
+
|
|
+ if (dev->oldest_dirty_seq) {
|
|
+ if (dev->oldest_dirty_seq > bi->seq_number) {
|
|
+ dev->oldest_dirty_seq = bi->seq_number;
|
|
+ dev->oldest_dirty_block = block_no;
|
|
+ }
|
|
+ }
|
|
+}
|
|
+
|
|
+int yaffs_block_ok_for_gc(struct yaffs_dev *dev, struct yaffs_block_info *bi)
|
|
+{
|
|
+
|
|
+ if (!dev->param.is_yaffs2)
|
|
+ return 1; /* disqualification only applies to yaffs2. */
|
|
+
|
|
+ if (!bi->has_shrink_hdr)
|
|
+ return 1; /* can gc */
|
|
+
|
|
+ yaffs2_find_oldest_dirty_seq(dev);
|
|
+
|
|
+ /* Can't do gc of this block if there are any blocks older than this
|
|
+ * one that have discarded pages.
|
|
+ */
|
|
+ return (bi->seq_number <= dev->oldest_dirty_seq);
|
|
+}
|
|
+
|
|
+/*
|
|
+ * yaffs2_find_refresh_block()
|
|
+ * periodically finds the oldest full block by sequence number for refreshing.
|
|
+ * Only for yaffs2.
|
|
+ */
|
|
+u32 yaffs2_find_refresh_block(struct yaffs_dev *dev)
|
|
+{
|
|
+ u32 b;
|
|
+ u32 oldest = 0;
|
|
+ u32 oldest_seq = 0;
|
|
+ struct yaffs_block_info *bi;
|
|
+
|
|
+ if (!dev->param.is_yaffs2)
|
|
+ return oldest;
|
|
+
|
|
+ /*
|
|
+ * If refresh period < 10 then refreshing is disabled.
|
|
+ */
|
|
+ if (dev->param.refresh_period < 10)
|
|
+ return oldest;
|
|
+
|
|
+ /*
|
|
+ * Fix broken values.
|
|
+ */
|
|
+ if (dev->refresh_skip > dev->param.refresh_period)
|
|
+ dev->refresh_skip = dev->param.refresh_period;
|
|
+
|
|
+ if (dev->refresh_skip > 0)
|
|
+ return oldest;
|
|
+
|
|
+ /*
|
|
+ * Refresh skip is now zero.
|
|
+ * We'll do a refresh this time around....
|
|
+ * Update the refresh skip and find the oldest block.
|
|
+ */
|
|
+ dev->refresh_skip = dev->param.refresh_period;
|
|
+ dev->refresh_count++;
|
|
+ bi = dev->block_info;
|
|
+ for (b = dev->internal_start_block; b <= dev->internal_end_block; b++) {
|
|
+
|
|
+ if (bi->block_state == YAFFS_BLOCK_STATE_FULL) {
|
|
+
|
|
+ if (oldest < 1 || bi->seq_number < oldest_seq) {
|
|
+ oldest = b;
|
|
+ oldest_seq = bi->seq_number;
|
|
+ }
|
|
+ }
|
|
+ bi++;
|
|
+ }
|
|
+
|
|
+ if (oldest > 0) {
|
|
+ yaffs_trace(YAFFS_TRACE_GC,
|
|
+ "GC refresh count %d selected block %d with seq_number %d",
|
|
+ dev->refresh_count, oldest, oldest_seq);
|
|
+ }
|
|
+
|
|
+ return oldest;
|
|
+}
|
|
+
|
|
+int yaffs2_checkpt_required(struct yaffs_dev *dev)
|
|
+{
|
|
+ int nblocks;
|
|
+
|
|
+ if (!dev->param.is_yaffs2)
|
|
+ return 0;
|
|
+
|
|
+ nblocks = dev->internal_end_block - dev->internal_start_block + 1;
|
|
+
|
|
+ return !dev->param.skip_checkpt_wr &&
|
|
+ !dev->read_only && (nblocks >= YAFFS_CHECKPOINT_MIN_BLOCKS);
|
|
+}
|
|
+
|
|
+int yaffs_calc_checkpt_blocks_required(struct yaffs_dev *dev)
|
|
+{
|
|
+ int retval;
|
|
+ int n_bytes = 0;
|
|
+ int n_blocks;
|
|
+ int dev_blocks;
|
|
+
|
|
+ if (!dev->param.is_yaffs2)
|
|
+ return 0;
|
|
+
|
|
+ if (!dev->checkpoint_blocks_required && yaffs2_checkpt_required(dev)) {
|
|
+ /* Not a valid value so recalculate */
|
|
+ dev_blocks = dev->param.end_block - dev->param.start_block + 1;
|
|
+ n_bytes += sizeof(struct yaffs_checkpt_validity);
|
|
+ n_bytes += sizeof(struct yaffs_checkpt_dev);
|
|
+ n_bytes += dev_blocks * sizeof(struct yaffs_block_info);
|
|
+ n_bytes += dev_blocks * dev->chunk_bit_stride;
|
|
+ n_bytes +=
|
|
+ (sizeof(struct yaffs_checkpt_obj) + sizeof(u32)) *
|
|
+ dev->n_obj;
|
|
+ n_bytes += (dev->tnode_size + sizeof(u32)) * dev->n_tnodes;
|
|
+ n_bytes += sizeof(struct yaffs_checkpt_validity);
|
|
+ n_bytes += sizeof(u32); /* checksum */
|
|
+
|
|
+ /* Round up and add 2 blocks to allow for some bad blocks,
|
|
+ * so add 3 */
|
|
+
|
|
+ n_blocks =
|
|
+ (n_bytes /
|
|
+ (dev->data_bytes_per_chunk *
|
|
+ dev->param.chunks_per_block)) + 3;
|
|
+
|
|
+ dev->checkpoint_blocks_required = n_blocks;
|
|
+ }
|
|
+
|
|
+ retval = dev->checkpoint_blocks_required - dev->blocks_in_checkpt;
|
|
+ if (retval < 0)
|
|
+ retval = 0;
|
|
+ return retval;
|
|
+}
|
|
+
|
|
+/*--------------------- Checkpointing --------------------*/
|
|
+
|
|
+static void yaffs2_do_endian_validity_marker(struct yaffs_dev *dev,
|
|
+ struct yaffs_checkpt_validity *v)
|
|
+{
|
|
+
|
|
+ if (!dev->swap_endian)
|
|
+ return;
|
|
+ v->struct_type = swap_s32(v->struct_type);
|
|
+ v->magic = swap_u32(v->magic);
|
|
+ v->version = swap_u32(v->version);
|
|
+ v->head = swap_u32(v->head);
|
|
+}
|
|
+
|
|
+static int yaffs2_wr_checkpt_validity_marker(struct yaffs_dev *dev, int head)
|
|
+{
|
|
+ struct yaffs_checkpt_validity cp;
|
|
+
|
|
+ memset(&cp, 0, sizeof(cp));
|
|
+
|
|
+ cp.struct_type = sizeof(cp);
|
|
+ cp.magic = YAFFS_MAGIC;
|
|
+ cp.version = YAFFS_CHECKPOINT_VERSION;
|
|
+ cp.head = (head) ? 1 : 0;
|
|
+
|
|
+ yaffs2_do_endian_validity_marker(dev, &cp);
|
|
+
|
|
+ return (yaffs2_checkpt_wr(dev, &cp, sizeof(cp)) == sizeof(cp)) ? 1 : 0;
|
|
+}
|
|
+
|
|
+static int yaffs2_rd_checkpt_validity_marker(struct yaffs_dev *dev, int head)
|
|
+{
|
|
+ struct yaffs_checkpt_validity cp;
|
|
+ int ok;
|
|
+
|
|
+ ok = (yaffs2_checkpt_rd(dev, &cp, sizeof(cp)) == sizeof(cp));
|
|
+ yaffs2_do_endian_validity_marker(dev, &cp);
|
|
+
|
|
+ if (ok)
|
|
+ ok = (cp.struct_type == sizeof(cp)) &&
|
|
+ (cp.magic == YAFFS_MAGIC) &&
|
|
+ (cp.version == YAFFS_CHECKPOINT_VERSION) &&
|
|
+ (cp.head == ((head) ? 1 : 0));
|
|
+ return ok ? 1 : 0;
|
|
+}
|
|
+
|
|
+static void yaffs2_dev_to_checkpt_dev(struct yaffs_checkpt_dev *cp,
|
|
+ struct yaffs_dev *dev)
|
|
+{
|
|
+ cp->struct_type = sizeof(*cp);
|
|
+
|
|
+ cp->n_erased_blocks = dev->n_erased_blocks;
|
|
+ cp->alloc_block = dev->alloc_block;
|
|
+ cp->alloc_page = dev->alloc_page;
|
|
+ cp->n_free_chunks = dev->n_free_chunks;
|
|
+
|
|
+ cp->n_deleted_files = dev->n_deleted_files;
|
|
+ cp->n_unlinked_files = dev->n_unlinked_files;
|
|
+ cp->n_bg_deletions = dev->n_bg_deletions;
|
|
+ cp->seq_number = dev->seq_number;
|
|
+
|
|
+}
|
|
+
|
|
+static void yaffs_checkpt_dev_to_dev(struct yaffs_dev *dev,
|
|
+ struct yaffs_checkpt_dev *cp)
|
|
+{
|
|
+ dev->n_erased_blocks = cp->n_erased_blocks;
|
|
+ dev->alloc_block = cp->alloc_block;
|
|
+ dev->alloc_page = cp->alloc_page;
|
|
+ dev->n_free_chunks = cp->n_free_chunks;
|
|
+
|
|
+ dev->n_deleted_files = cp->n_deleted_files;
|
|
+ dev->n_unlinked_files = cp->n_unlinked_files;
|
|
+ dev->n_bg_deletions = cp->n_bg_deletions;
|
|
+ dev->seq_number = cp->seq_number;
|
|
+}
|
|
+
|
|
+static void yaffs2_do_endian_checkpt_dev(struct yaffs_dev *dev,
|
|
+ struct yaffs_checkpt_dev *cp)
|
|
+{
|
|
+ if (!dev->swap_endian)
|
|
+ return;
|
|
+ cp->struct_type = swap_s32(cp->struct_type);
|
|
+ cp->n_erased_blocks = swap_s32(cp->n_erased_blocks);
|
|
+ cp->alloc_block = swap_s32(cp->alloc_block);
|
|
+ cp->alloc_page = swap_u32(cp->alloc_page);
|
|
+ cp->n_free_chunks = swap_s32(cp->n_free_chunks);
|
|
+ cp->n_deleted_files = swap_s32(cp->n_deleted_files);
|
|
+ cp->n_unlinked_files = swap_s32(cp->n_unlinked_files);
|
|
+ cp->n_bg_deletions = swap_s32(cp->n_bg_deletions);
|
|
+}
|
|
+
|
|
+static int yaffs2_wr_checkpt_dev(struct yaffs_dev *dev)
|
|
+{
|
|
+ struct yaffs_checkpt_dev cp;
|
|
+ u32 n_bytes;
|
|
+ u32 n_blocks = dev->internal_end_block - dev->internal_start_block + 1;
|
|
+ int ok;
|
|
+ u32 i;
|
|
+ union yaffs_block_info_union bu;
|
|
+
|
|
+ /* Write device runtime values */
|
|
+ yaffs2_dev_to_checkpt_dev(&cp, dev);
|
|
+ yaffs2_do_endian_checkpt_dev(dev, &cp);
|
|
+
|
|
+ ok = (yaffs2_checkpt_wr(dev, &cp, sizeof(cp)) == sizeof(cp));
|
|
+ if (!ok)
|
|
+ return 0;
|
|
+
|
|
+ /* Write block info. */
|
|
+ if (!dev->swap_endian) {
|
|
+ n_bytes = n_blocks * sizeof(struct yaffs_block_info);
|
|
+ ok = (yaffs2_checkpt_wr(dev, dev->block_info, n_bytes) ==
|
|
+ (int)n_bytes);
|
|
+ } else {
|
|
+ /*
|
|
+ * Need to swap the endianisms. We can't do this in place
|
|
+ * since that would damage live data,
|
|
+ * so write one block info at a time using a copy.
|
|
+ */
|
|
+ for (i = 0; i < n_blocks && ok; i++) {
|
|
+ bu.bi = dev->block_info[i];
|
|
+ bu.as_u32[0] = swap_u32(bu.as_u32[0]);
|
|
+ bu.as_u32[1] = swap_u32(bu.as_u32[1]);
|
|
+ ok = (yaffs2_checkpt_wr(dev, &bu, sizeof(bu)) == sizeof(bu));
|
|
+ }
|
|
+ }
|
|
+
|
|
+ if (!ok)
|
|
+ return 0;
|
|
+
|
|
+ /*
|
|
+ * Write chunk bits. Chunk bits are in bytes so
|
|
+ * no endian conversion is needed.
|
|
+ */
|
|
+ n_bytes = n_blocks * dev->chunk_bit_stride;
|
|
+ ok = (yaffs2_checkpt_wr(dev, dev->chunk_bits, n_bytes) ==
|
|
+ (int)n_bytes);
|
|
+
|
|
+ return ok ? 1 : 0;
|
|
+}
|
|
+
|
|
+static int yaffs2_rd_checkpt_dev(struct yaffs_dev *dev)
|
|
+{
|
|
+ struct yaffs_checkpt_dev cp;
|
|
+ u32 n_bytes;
|
|
+ u32 n_blocks =
|
|
+ (dev->internal_end_block - dev->internal_start_block + 1);
|
|
+ int ok;
|
|
+
|
|
+ ok = (yaffs2_checkpt_rd(dev, &cp, sizeof(cp)) == sizeof(cp));
|
|
+ if (!ok)
|
|
+ return 0;
|
|
+ yaffs2_do_endian_checkpt_dev(dev, &cp);
|
|
+
|
|
+ if (cp.struct_type != sizeof(cp))
|
|
+ return 0;
|
|
+
|
|
+ yaffs_checkpt_dev_to_dev(dev, &cp);
|
|
+
|
|
+ n_bytes = n_blocks * sizeof(struct yaffs_block_info);
|
|
+
|
|
+ ok = (yaffs2_checkpt_rd(dev, dev->block_info, n_bytes) ==
|
|
+ (int)n_bytes);
|
|
+
|
|
+ if (!ok)
|
|
+ return 0;
|
|
+
|
|
+ if (dev->swap_endian) {
|
|
+ /* The block info can just be handled as a list of u32s. */
|
|
+ u32 *as_u32 = (u32 *) dev->block_info;
|
|
+ u32 n_u32s = n_bytes/sizeof(u32);
|
|
+ u32 i;
|
|
+
|
|
+ for (i=0; i < n_u32s; i++)
|
|
+ as_u32[i] = swap_u32(as_u32[i]);
|
|
+ }
|
|
+
|
|
+ n_bytes = n_blocks * dev->chunk_bit_stride;
|
|
+
|
|
+ ok = (yaffs2_checkpt_rd(dev, dev->chunk_bits, n_bytes) ==
|
|
+ (int)n_bytes);
|
|
+
|
|
+
|
|
+ return ok ? 1 : 0;
|
|
+}
|
|
+
|
|
+
|
|
+static void yaffs2_checkpt_obj_bit_assign(struct yaffs_checkpt_obj *cp,
|
|
+ int bit_offset,
|
|
+ int bit_width,
|
|
+ u32 value)
|
|
+{
|
|
+ u32 and_mask;
|
|
+
|
|
+ and_mask = ((1<<bit_width)-1) << bit_offset;
|
|
+
|
|
+ cp->bit_field &= ~and_mask;
|
|
+ cp->bit_field |= ((value << bit_offset) & and_mask);
|
|
+}
|
|
+
|
|
+static u32 yaffs2_checkpt_obj_bit_get(struct yaffs_checkpt_obj *cp,
|
|
+ int bit_offset,
|
|
+ int bit_width)
|
|
+{
|
|
+ u32 and_mask;
|
|
+
|
|
+ and_mask = ((1<<bit_width)-1);
|
|
+
|
|
+ return (cp->bit_field >> bit_offset) & and_mask;
|
|
+}
|
|
+
|
|
+static void yaffs2_obj_checkpt_obj(struct yaffs_checkpt_obj *cp,
|
|
+ struct yaffs_obj *obj)
|
|
+{
|
|
+ cp->obj_id = obj->obj_id;
|
|
+ cp->parent_id = (obj->parent) ? obj->parent->obj_id : 0;
|
|
+ cp->hdr_chunk = obj->hdr_chunk;
|
|
+
|
|
+ yaffs2_checkpt_obj_bit_assign(cp, CHECKPOINT_VARIANT_BITS, obj->variant_type);
|
|
+ yaffs2_checkpt_obj_bit_assign(cp, CHECKPOINT_DELETED_BITS, obj->deleted);
|
|
+ yaffs2_checkpt_obj_bit_assign(cp, CHECKPOINT_SOFT_DEL_BITS, obj->soft_del);
|
|
+ yaffs2_checkpt_obj_bit_assign(cp, CHECKPOINT_UNLINKED_BITS, obj->unlinked);
|
|
+ yaffs2_checkpt_obj_bit_assign(cp, CHECKPOINT_FAKE_BITS, obj->fake);
|
|
+ yaffs2_checkpt_obj_bit_assign(cp, CHECKPOINT_RENAME_ALLOWED_BITS, obj->rename_allowed);
|
|
+ yaffs2_checkpt_obj_bit_assign(cp, CHECKPOINT_UNLINK_ALLOWED_BITS, obj->unlink_allowed);
|
|
+ yaffs2_checkpt_obj_bit_assign(cp, CHECKPOINT_SERIAL_BITS, obj->serial);
|
|
+
|
|
+ cp->n_data_chunks = obj->n_data_chunks;
|
|
+
|
|
+ if (obj->variant_type == YAFFS_OBJECT_TYPE_FILE)
|
|
+ cp->size_or_equiv_obj = obj->variant.file_variant.file_size;
|
|
+ else if (obj->variant_type == YAFFS_OBJECT_TYPE_HARDLINK)
|
|
+ cp->size_or_equiv_obj = obj->variant.hardlink_variant.equiv_id;
|
|
+}
|
|
+
|
|
+static int yaffs2_checkpt_obj_to_obj(struct yaffs_obj *obj,
|
|
+ struct yaffs_checkpt_obj *cp)
|
|
+{
|
|
+ struct yaffs_obj *parent;
|
|
+ u32 cp_variant_type = yaffs2_checkpt_obj_bit_get(cp, CHECKPOINT_VARIANT_BITS);
|
|
+
|
|
+ if (obj->variant_type != cp_variant_type) {
|
|
+ yaffs_trace(YAFFS_TRACE_ERROR,
|
|
+ "Checkpoint read object %d type %d chunk %d does not match existing object type %d",
|
|
+ cp->obj_id, cp_variant_type, cp->hdr_chunk,
|
|
+ obj->variant_type);
|
|
+ return 0;
|
|
+ }
|
|
+
|
|
+ obj->obj_id = cp->obj_id;
|
|
+
|
|
+ if (cp->parent_id)
|
|
+ parent = yaffs_find_or_create_by_number(obj->my_dev,
|
|
+ cp->parent_id,
|
|
+ YAFFS_OBJECT_TYPE_DIRECTORY);
|
|
+ else
|
|
+ parent = NULL;
|
|
+
|
|
+ if (parent) {
|
|
+ if (parent->variant_type != YAFFS_OBJECT_TYPE_DIRECTORY) {
|
|
+ yaffs_trace(YAFFS_TRACE_ALWAYS,
|
|
+ "Checkpoint read object %d parent %d type %d chunk %d Parent type, %d, not directory",
|
|
+ cp->obj_id, cp->parent_id,
|
|
+ cp_variant_type, cp->hdr_chunk,
|
|
+ parent->variant_type);
|
|
+ return 0;
|
|
+ }
|
|
+ yaffs_add_obj_to_dir(parent, obj);
|
|
+ }
|
|
+
|
|
+ obj->hdr_chunk = cp->hdr_chunk;
|
|
+
|
|
+ obj->variant_type = yaffs2_checkpt_obj_bit_get(cp, CHECKPOINT_VARIANT_BITS);
|
|
+ obj->deleted = yaffs2_checkpt_obj_bit_get(cp, CHECKPOINT_DELETED_BITS);
|
|
+ obj->soft_del = yaffs2_checkpt_obj_bit_get(cp, CHECKPOINT_SOFT_DEL_BITS);
|
|
+ obj->unlinked = yaffs2_checkpt_obj_bit_get(cp, CHECKPOINT_UNLINKED_BITS);
|
|
+ obj->fake = yaffs2_checkpt_obj_bit_get(cp, CHECKPOINT_FAKE_BITS);
|
|
+ obj->rename_allowed = yaffs2_checkpt_obj_bit_get(cp, CHECKPOINT_RENAME_ALLOWED_BITS);
|
|
+ obj->unlink_allowed = yaffs2_checkpt_obj_bit_get(cp, CHECKPOINT_UNLINK_ALLOWED_BITS);
|
|
+ obj->serial = yaffs2_checkpt_obj_bit_get(cp, CHECKPOINT_SERIAL_BITS);
|
|
+
|
|
+ obj->n_data_chunks = cp->n_data_chunks;
|
|
+
|
|
+ if (obj->variant_type == YAFFS_OBJECT_TYPE_FILE) {
|
|
+ obj->variant.file_variant.file_size = cp->size_or_equiv_obj;
|
|
+ obj->variant.file_variant.stored_size = cp->size_or_equiv_obj;
|
|
+ } else if (obj->variant_type == YAFFS_OBJECT_TYPE_HARDLINK) {
|
|
+ obj->variant.hardlink_variant.equiv_id = cp->size_or_equiv_obj;
|
|
+ }
|
|
+ if (obj->hdr_chunk > 0)
|
|
+ obj->lazy_loaded = 1;
|
|
+ return 1;
|
|
+}
|
|
+
|
|
+static void yaffs2_do_endian_tnode(struct yaffs_dev *dev, struct yaffs_tnode *tn)
|
|
+{
|
|
+ int i;
|
|
+ u32 *as_u32 = (u32 *)tn;
|
|
+ int tnode_size_u32 = dev->tnode_size / sizeof(u32);
|
|
+
|
|
+ if (!dev->swap_endian)
|
|
+ return;
|
|
+ /* Swap all the tnode data as u32s to fix endianisms. */
|
|
+ for (i = 0; i<tnode_size_u32; i++)
|
|
+ as_u32[i] = swap_u32(as_u32[i]);
|
|
+}
|
|
+
|
|
+struct yaffs_tnode *yaffs2_do_endian_tnode_copy(struct yaffs_dev *dev,
|
|
+ struct yaffs_tnode *tn)
|
|
+{
|
|
+ if (!dev->swap_endian)
|
|
+ return tn;
|
|
+
|
|
+ memcpy(dev->tn_swap_buffer, tn, dev->tnode_size);
|
|
+ tn = dev->tn_swap_buffer;
|
|
+
|
|
+ yaffs2_do_endian_tnode(dev, tn);
|
|
+
|
|
+ return tn;
|
|
+}
|
|
+
|
|
+static int yaffs2_checkpt_tnode_worker(struct yaffs_obj *in,
|
|
+ struct yaffs_tnode *tn, u32 level,
|
|
+ int chunk_offset)
|
|
+{
|
|
+ int i;
|
|
+ struct yaffs_dev *dev = in->my_dev;
|
|
+ int ok = 1;
|
|
+ u32 base_offset;
|
|
+
|
|
+ if (!tn)
|
|
+ return 1;
|
|
+
|
|
+ if (level > 0) {
|
|
+ for (i = 0; i < YAFFS_NTNODES_INTERNAL && ok; i++) {
|
|
+ if (!tn->internal[i])
|
|
+ continue;
|
|
+ ok = yaffs2_checkpt_tnode_worker(in,
|
|
+ tn->internal[i],
|
|
+ level - 1,
|
|
+ (chunk_offset <<
|
|
+ YAFFS_TNODES_INTERNAL_BITS) + i);
|
|
+ }
|
|
+ return ok;
|
|
+ }
|
|
+
|
|
+ /* Level 0 tnode */
|
|
+ base_offset = chunk_offset << YAFFS_TNODES_LEVEL0_BITS;
|
|
+ yaffs_do_endian_u32(dev, &base_offset);
|
|
+
|
|
+ ok = (yaffs2_checkpt_wr(dev, &base_offset, sizeof(base_offset)) ==
|
|
+ sizeof(base_offset));
|
|
+ if (ok) {
|
|
+ /*
|
|
+ * NB Can't do an in-place endian swizzle since that would
|
|
+ * damage current tnode data.
|
|
+ * If a tnode endian conversion is required we do a copy.
|
|
+ */
|
|
+ tn = yaffs2_do_endian_tnode_copy(dev, tn);
|
|
+ ok = (yaffs2_checkpt_wr(dev, tn, dev->tnode_size) ==
|
|
+ (int)dev->tnode_size);
|
|
+ }
|
|
+ return ok;
|
|
+}
|
|
+
|
|
+static int yaffs2_wr_checkpt_tnodes(struct yaffs_obj *obj)
|
|
+{
|
|
+ u32 end_marker = ~0;
|
|
+ int ok = 1;
|
|
+
|
|
+ if (obj->variant_type != YAFFS_OBJECT_TYPE_FILE)
|
|
+ return ok;
|
|
+
|
|
+ ok = yaffs2_checkpt_tnode_worker(obj,
|
|
+ obj->variant.file_variant.top,
|
|
+ obj->variant.file_variant.
|
|
+ top_level, 0);
|
|
+ if (ok)
|
|
+ ok = (yaffs2_checkpt_wr(obj->my_dev, &end_marker,
|
|
+ sizeof(end_marker)) == sizeof(end_marker));
|
|
+
|
|
+ return ok ? 1 : 0;
|
|
+}
|
|
+
|
|
+static int yaffs2_rd_checkpt_tnodes(struct yaffs_obj *obj)
|
|
+{
|
|
+ u32 base_chunk;
|
|
+ int ok = 1;
|
|
+ struct yaffs_dev *dev = obj->my_dev;
|
|
+ struct yaffs_file_var *file_stuct_ptr = &obj->variant.file_variant;
|
|
+ struct yaffs_tnode *tn;
|
|
+ int nread = 0;
|
|
+
|
|
+ ok = (yaffs2_checkpt_rd(dev, &base_chunk, sizeof(base_chunk)) ==
|
|
+ sizeof(base_chunk));
|
|
+
|
|
+ yaffs_do_endian_u32(dev, &base_chunk);
|
|
+
|
|
+ while (ok && (~base_chunk)) {
|
|
+ nread++;
|
|
+ /* Read level 0 tnode */
|
|
+
|
|
+ tn = yaffs_get_tnode(dev);
|
|
+ if (tn) {
|
|
+ ok = (yaffs2_checkpt_rd(dev, tn, dev->tnode_size) ==
|
|
+ (int)dev->tnode_size);
|
|
+ yaffs2_do_endian_tnode(dev, tn);
|
|
+ }
|
|
+ else
|
|
+ ok = 0;
|
|
+
|
|
+ if (tn && ok)
|
|
+ ok = yaffs_add_find_tnode_0(dev,
|
|
+ file_stuct_ptr,
|
|
+ base_chunk, tn) ? 1 : 0;
|
|
+
|
|
+ if (ok) {
|
|
+ ok = (yaffs2_checkpt_rd
|
|
+ (dev, &base_chunk,
|
|
+ sizeof(base_chunk)) == sizeof(base_chunk));
|
|
+ yaffs_do_endian_u32(dev, &base_chunk);
|
|
+ }
|
|
+
|
|
+ }
|
|
+
|
|
+ yaffs_trace(YAFFS_TRACE_CHECKPOINT,
|
|
+ "Checkpoint read tnodes %d records, last %d. ok %d",
|
|
+ nread, base_chunk, ok);
|
|
+
|
|
+ return ok ? 1 : 0;
|
|
+}
|
|
+
|
|
+
|
|
+static void yaffs2_do_endian_checkpt_obj(struct yaffs_dev *dev,
|
|
+ struct yaffs_checkpt_obj *cp)
|
|
+{
|
|
+ if (!dev->swap_endian)
|
|
+ return;
|
|
+ cp->struct_type = swap_s32(cp->struct_type);
|
|
+ cp->obj_id = swap_u32(cp->obj_id);
|
|
+ cp->parent_id = swap_u32(cp->parent_id);
|
|
+ cp->hdr_chunk = swap_s32(cp->hdr_chunk);
|
|
+ cp->bit_field = swap_u32(cp->bit_field);
|
|
+ cp->n_data_chunks = swap_s32(cp->n_data_chunks);
|
|
+ cp->size_or_equiv_obj = swap_loff_t(cp->size_or_equiv_obj);
|
|
+}
|
|
+
|
|
+static int yaffs2_wr_checkpt_objs(struct yaffs_dev *dev)
|
|
+{
|
|
+ struct yaffs_obj *obj;
|
|
+ struct yaffs_checkpt_obj cp;
|
|
+ int i;
|
|
+ int ok = 1;
|
|
+ struct list_head *lh;
|
|
+ u32 cp_variant_type;
|
|
+
|
|
+ /* Iterate through the objects in each hash entry,
|
|
+ * dumping them to the checkpointing stream.
|
|
+ */
|
|
+
|
|
+ (void) cp_variant_type;
|
|
+
|
|
+ for (i = 0; ok && i < YAFFS_NOBJECT_BUCKETS; i++) {
|
|
+ list_for_each(lh, &dev->obj_bucket[i].list) {
|
|
+ obj = list_entry(lh, struct yaffs_obj, hash_link);
|
|
+ if (!obj->defered_free) {
|
|
+ yaffs2_obj_checkpt_obj(&cp, obj);
|
|
+ cp.struct_type = sizeof(cp);
|
|
+ cp_variant_type = yaffs2_checkpt_obj_bit_get(
|
|
+ &cp, CHECKPOINT_VARIANT_BITS);
|
|
+ yaffs_trace(YAFFS_TRACE_CHECKPOINT,
|
|
+ "Checkpoint write object %d parent %d type %d chunk %d obj addr %p",
|
|
+ cp.obj_id, cp.parent_id,
|
|
+ cp_variant_type, cp.hdr_chunk, obj);
|
|
+
|
|
+ yaffs2_do_endian_checkpt_obj (dev, &cp);
|
|
+ ok = (yaffs2_checkpt_wr(dev, &cp,
|
|
+ sizeof(cp)) == sizeof(cp));
|
|
+
|
|
+ if (ok &&
|
|
+ obj->variant_type ==
|
|
+ YAFFS_OBJECT_TYPE_FILE)
|
|
+ ok = yaffs2_wr_checkpt_tnodes(obj);
|
|
+ }
|
|
+ }
|
|
+ }
|
|
+
|
|
+ /* Dump end of list */
|
|
+ memset(&cp, 0xff, sizeof(struct yaffs_checkpt_obj));
|
|
+ cp.struct_type = sizeof(cp);
|
|
+ yaffs2_do_endian_checkpt_obj (dev, &cp);
|
|
+
|
|
+ if (ok)
|
|
+ ok = (yaffs2_checkpt_wr(dev, &cp, sizeof(cp)) == sizeof(cp));
|
|
+
|
|
+ return ok ? 1 : 0;
|
|
+}
|
|
+
|
|
+static int yaffs2_rd_checkpt_objs(struct yaffs_dev *dev)
|
|
+{
|
|
+ struct yaffs_obj *obj;
|
|
+ struct yaffs_checkpt_obj cp;
|
|
+ int ok = 1;
|
|
+ int done = 0;
|
|
+ u32 cp_variant_type;
|
|
+ LIST_HEAD(hard_list);
|
|
+
|
|
+
|
|
+ while (ok && !done) {
|
|
+ ok = (yaffs2_checkpt_rd(dev, &cp, sizeof(cp)) == sizeof(cp));
|
|
+ yaffs2_do_endian_checkpt_obj (dev, &cp);
|
|
+
|
|
+ if (cp.struct_type != sizeof(cp)) {
|
|
+ yaffs_trace(YAFFS_TRACE_CHECKPOINT,
|
|
+ "struct size %d instead of %d ok %d",
|
|
+ cp.struct_type, (int)sizeof(cp), ok);
|
|
+ ok = 0;
|
|
+ }
|
|
+
|
|
+ cp_variant_type = yaffs2_checkpt_obj_bit_get(
|
|
+ &cp, CHECKPOINT_VARIANT_BITS);
|
|
+ yaffs_trace(YAFFS_TRACE_CHECKPOINT,
|
|
+ "Checkpoint read object %d parent %d type %d chunk %d ",
|
|
+ cp.obj_id, cp.parent_id, cp_variant_type,
|
|
+ cp.hdr_chunk);
|
|
+
|
|
+ if (ok && cp.obj_id == (u32)(~0)) {
|
|
+ done = 1;
|
|
+ } else if (ok) {
|
|
+ obj =
|
|
+ yaffs_find_or_create_by_number(dev, cp.obj_id,
|
|
+ cp_variant_type);
|
|
+ if (obj) {
|
|
+ ok = yaffs2_checkpt_obj_to_obj(obj, &cp);
|
|
+ if (!ok)
|
|
+ break;
|
|
+ if (obj->variant_type ==
|
|
+ YAFFS_OBJECT_TYPE_FILE) {
|
|
+ ok = yaffs2_rd_checkpt_tnodes(obj);
|
|
+ } else if (obj->variant_type ==
|
|
+ YAFFS_OBJECT_TYPE_HARDLINK) {
|
|
+ list_add(&obj->hard_links, &hard_list);
|
|
+ }
|
|
+ } else {
|
|
+ ok = 0;
|
|
+ }
|
|
+ }
|
|
+ }
|
|
+
|
|
+ if (ok)
|
|
+ yaffs_link_fixup(dev, &hard_list);
|
|
+
|
|
+ return ok ? 1 : 0;
|
|
+}
|
|
+
|
|
+static int yaffs2_wr_checkpt_sum(struct yaffs_dev *dev)
|
|
+{
|
|
+ u32 checkpt_sum;
|
|
+ int ok;
|
|
+
|
|
+ yaffs2_get_checkpt_sum(dev, &checkpt_sum);
|
|
+
|
|
+ yaffs_do_endian_u32(dev, &checkpt_sum);
|
|
+
|
|
+ ok = (yaffs2_checkpt_wr(dev, &checkpt_sum, sizeof(checkpt_sum)) ==
|
|
+ sizeof(checkpt_sum));
|
|
+
|
|
+ if (!ok)
|
|
+ return 0;
|
|
+
|
|
+ return 1;
|
|
+}
|
|
+
|
|
+static int yaffs2_rd_checkpt_sum(struct yaffs_dev *dev)
|
|
+{
|
|
+ u32 checkpt_sum0;
|
|
+ u32 checkpt_sum1;
|
|
+ int ok;
|
|
+
|
|
+ yaffs2_get_checkpt_sum(dev, &checkpt_sum0);
|
|
+
|
|
+ ok = (yaffs2_checkpt_rd(dev, &checkpt_sum1, sizeof(checkpt_sum1)) ==
|
|
+ sizeof(checkpt_sum1));
|
|
+
|
|
+ if (!ok)
|
|
+ return 0;
|
|
+ yaffs_do_endian_u32(dev, &checkpt_sum1);
|
|
+
|
|
+ if (checkpt_sum0 != checkpt_sum1)
|
|
+ return 0;
|
|
+
|
|
+ return 1;
|
|
+}
|
|
+
|
|
+static int yaffs2_wr_checkpt_data(struct yaffs_dev *dev)
|
|
+{
|
|
+ int ok = 1;
|
|
+
|
|
+ if (!yaffs2_checkpt_required(dev)) {
|
|
+ yaffs_trace(YAFFS_TRACE_CHECKPOINT,
|
|
+ "skipping checkpoint write");
|
|
+ ok = 0;
|
|
+ }
|
|
+
|
|
+ if (ok)
|
|
+ ok = yaffs2_checkpt_open(dev, 1);
|
|
+
|
|
+ if (ok) {
|
|
+ yaffs_trace(YAFFS_TRACE_CHECKPOINT,
|
|
+ "write checkpoint validity");
|
|
+ ok = yaffs2_wr_checkpt_validity_marker(dev, 1);
|
|
+ }
|
|
+ if (ok) {
|
|
+ yaffs_trace(YAFFS_TRACE_CHECKPOINT,
|
|
+ "write checkpoint device");
|
|
+ ok = yaffs2_wr_checkpt_dev(dev);
|
|
+ }
|
|
+ if (ok) {
|
|
+ yaffs_trace(YAFFS_TRACE_CHECKPOINT,
|
|
+ "write checkpoint objects");
|
|
+ ok = yaffs2_wr_checkpt_objs(dev);
|
|
+ }
|
|
+ if (ok) {
|
|
+ yaffs_trace(YAFFS_TRACE_CHECKPOINT,
|
|
+ "write checkpoint validity");
|
|
+ ok = yaffs2_wr_checkpt_validity_marker(dev, 0);
|
|
+ }
|
|
+
|
|
+ if (ok)
|
|
+ ok = yaffs2_wr_checkpt_sum(dev);
|
|
+
|
|
+ if (!yaffs_checkpt_close(dev))
|
|
+ ok = 0;
|
|
+
|
|
+ if (ok)
|
|
+ dev->is_checkpointed = 1;
|
|
+ else
|
|
+ dev->is_checkpointed = 0;
|
|
+
|
|
+ return dev->is_checkpointed;
|
|
+}
|
|
+
|
|
+static int yaffs2_rd_checkpt_data(struct yaffs_dev *dev)
|
|
+{
|
|
+ int ok = 1;
|
|
+
|
|
+ if (!dev->param.is_yaffs2)
|
|
+ ok = 0;
|
|
+
|
|
+ if (ok && dev->param.skip_checkpt_rd) {
|
|
+ yaffs_trace(YAFFS_TRACE_CHECKPOINT,
|
|
+ "skipping checkpoint read");
|
|
+ ok = 0;
|
|
+ }
|
|
+
|
|
+ if (ok)
|
|
+ ok = yaffs2_checkpt_open(dev, 0); /* open for read */
|
|
+
|
|
+ if (ok) {
|
|
+ yaffs_trace(YAFFS_TRACE_CHECKPOINT,
|
|
+ "read checkpoint validity");
|
|
+ ok = yaffs2_rd_checkpt_validity_marker(dev, 1);
|
|
+ }
|
|
+ if (ok) {
|
|
+ yaffs_trace(YAFFS_TRACE_CHECKPOINT,
|
|
+ "read checkpoint device");
|
|
+ ok = yaffs2_rd_checkpt_dev(dev);
|
|
+ }
|
|
+ if (ok) {
|
|
+ yaffs_trace(YAFFS_TRACE_CHECKPOINT,
|
|
+ "read checkpoint objects");
|
|
+ ok = yaffs2_rd_checkpt_objs(dev);
|
|
+ }
|
|
+ if (ok) {
|
|
+ yaffs_trace(YAFFS_TRACE_CHECKPOINT,
|
|
+ "read checkpoint validity");
|
|
+ ok = yaffs2_rd_checkpt_validity_marker(dev, 0);
|
|
+ }
|
|
+
|
|
+ if (ok) {
|
|
+ ok = yaffs2_rd_checkpt_sum(dev);
|
|
+ yaffs_trace(YAFFS_TRACE_CHECKPOINT,
|
|
+ "read checkpoint checksum %d", ok);
|
|
+ }
|
|
+
|
|
+ if (!yaffs_checkpt_close(dev))
|
|
+ ok = 0;
|
|
+
|
|
+ if (ok)
|
|
+ dev->is_checkpointed = 1;
|
|
+ else
|
|
+ dev->is_checkpointed = 0;
|
|
+
|
|
+ return ok ? 1 : 0;
|
|
+}
|
|
+
|
|
+void yaffs2_checkpt_invalidate(struct yaffs_dev *dev)
|
|
+{
|
|
+ if (dev->is_checkpointed || dev->blocks_in_checkpt > 0) {
|
|
+ dev->is_checkpointed = 0;
|
|
+ yaffs2_checkpt_invalidate_stream(dev);
|
|
+ }
|
|
+ if (dev->param.sb_dirty_fn)
|
|
+ dev->param.sb_dirty_fn(dev);
|
|
+}
|
|
+
|
|
+int yaffs_checkpoint_save(struct yaffs_dev *dev)
|
|
+{
|
|
+ yaffs_trace(YAFFS_TRACE_CHECKPOINT,
|
|
+ "save entry: is_checkpointed %d",
|
|
+ dev->is_checkpointed);
|
|
+
|
|
+ yaffs_verify_objects(dev);
|
|
+ yaffs_verify_blocks(dev);
|
|
+ yaffs_verify_free_chunks(dev);
|
|
+
|
|
+ if (!dev->is_checkpointed) {
|
|
+ yaffs2_checkpt_invalidate(dev);
|
|
+ yaffs2_wr_checkpt_data(dev);
|
|
+ }
|
|
+
|
|
+ yaffs_trace(YAFFS_TRACE_CHECKPOINT | YAFFS_TRACE_MOUNT,
|
|
+ "save exit: is_checkpointed %d",
|
|
+ dev->is_checkpointed);
|
|
+
|
|
+ return dev->is_checkpointed;
|
|
+}
|
|
+
|
|
+int yaffs2_checkpt_restore(struct yaffs_dev *dev)
|
|
+{
|
|
+ int retval;
|
|
+
|
|
+ yaffs_trace(YAFFS_TRACE_CHECKPOINT,
|
|
+ "restore entry: is_checkpointed %d",
|
|
+ dev->is_checkpointed);
|
|
+
|
|
+ retval = yaffs2_rd_checkpt_data(dev);
|
|
+
|
|
+ if (dev->is_checkpointed) {
|
|
+ yaffs_verify_objects(dev);
|
|
+ yaffs_verify_blocks(dev);
|
|
+ yaffs_verify_free_chunks(dev);
|
|
+ }
|
|
+
|
|
+ yaffs_trace(YAFFS_TRACE_CHECKPOINT,
|
|
+ "restore exit: is_checkpointed %d",
|
|
+ dev->is_checkpointed);
|
|
+
|
|
+ return retval;
|
|
+}
|
|
+
|
|
+/* End of checkpointing */
|
|
+
|
|
+/* Hole handling logic for truncate past end of file */
|
|
+
|
|
+int yaffs2_handle_hole(struct yaffs_obj *obj, loff_t new_size)
|
|
+{
|
|
+ /* if new_size > old_file_size.
|
|
+ * We're going to be writing a hole.
|
|
+ * If the hole is small then write zeros otherwise write a start
|
|
+ * of hole marker.
|
|
+ */
|
|
+ loff_t old_file_size;
|
|
+ loff_t increase;
|
|
+ int small_hole;
|
|
+ int result = YAFFS_OK;
|
|
+ struct yaffs_dev *dev = NULL;
|
|
+ u8 *local_buffer = NULL;
|
|
+ int small_increase_ok = 0;
|
|
+
|
|
+ if (!obj)
|
|
+ return YAFFS_FAIL;
|
|
+
|
|
+ if (obj->variant_type != YAFFS_OBJECT_TYPE_FILE)
|
|
+ return YAFFS_FAIL;
|
|
+
|
|
+ dev = obj->my_dev;
|
|
+
|
|
+ /* Bail out if not yaffs2 mode */
|
|
+ if (!dev->param.is_yaffs2)
|
|
+ return YAFFS_OK;
|
|
+
|
|
+ old_file_size = obj->variant.file_variant.file_size;
|
|
+
|
|
+ if (new_size <= old_file_size)
|
|
+ return YAFFS_OK;
|
|
+
|
|
+ increase = new_size - old_file_size;
|
|
+
|
|
+ if (increase < YAFFS_SMALL_HOLE_THRESHOLD * dev->data_bytes_per_chunk &&
|
|
+ yaffs_check_alloc_available(dev, YAFFS_SMALL_HOLE_THRESHOLD + 1))
|
|
+ small_hole = 1;
|
|
+ else
|
|
+ small_hole = 0;
|
|
+
|
|
+ if (small_hole)
|
|
+ local_buffer = yaffs_get_temp_buffer(dev);
|
|
+
|
|
+ if (local_buffer) {
|
|
+ /* fill hole with zero bytes */
|
|
+ loff_t pos = old_file_size;
|
|
+ int this_write;
|
|
+ int written;
|
|
+ memset(local_buffer, 0, dev->data_bytes_per_chunk);
|
|
+ small_increase_ok = 1;
|
|
+
|
|
+ while (increase > 0 && small_increase_ok) {
|
|
+ this_write = increase;
|
|
+ if (this_write > (int)dev->data_bytes_per_chunk)
|
|
+ this_write = dev->data_bytes_per_chunk;
|
|
+ written =
|
|
+ yaffs_do_file_wr(obj, local_buffer, pos, this_write,
|
|
+ 0);
|
|
+ if (written == this_write) {
|
|
+ pos += this_write;
|
|
+ increase -= this_write;
|
|
+ } else {
|
|
+ small_increase_ok = 0;
|
|
+ }
|
|
+ }
|
|
+
|
|
+ yaffs_release_temp_buffer(dev, local_buffer);
|
|
+
|
|
+ /* If out of space then reverse any chunks we've added */
|
|
+ if (!small_increase_ok)
|
|
+ yaffs_resize_file_down(obj, old_file_size);
|
|
+ }
|
|
+
|
|
+ if (!small_increase_ok &&
|
|
+ obj->parent &&
|
|
+ obj->parent->obj_id != YAFFS_OBJECTID_UNLINKED &&
|
|
+ obj->parent->obj_id != YAFFS_OBJECTID_DELETED) {
|
|
+ /* Write a hole start header with the old file size */
|
|
+ yaffs_update_oh(obj, NULL, 0, 1, 0, NULL);
|
|
+ }
|
|
+
|
|
+ return result;
|
|
+}
|
|
+
|
|
+/* Yaffs2 scanning */
|
|
+
|
|
+struct yaffs_block_index {
|
|
+ int seq;
|
|
+ int block;
|
|
+};
|
|
+
|
|
+static int yaffs2_ybicmp(const void *a, const void *b)
|
|
+{
|
|
+ int aseq = ((struct yaffs_block_index *)a)->seq;
|
|
+ int bseq = ((struct yaffs_block_index *)b)->seq;
|
|
+ int ablock = ((struct yaffs_block_index *)a)->block;
|
|
+ int bblock = ((struct yaffs_block_index *)b)->block;
|
|
+
|
|
+ if (aseq == bseq)
|
|
+ return ablock - bblock;
|
|
+
|
|
+ return aseq - bseq;
|
|
+}
|
|
+
|
|
+static inline int yaffs2_scan_chunk(struct yaffs_dev *dev,
|
|
+ struct yaffs_block_info *bi,
|
|
+ int blk, int chunk_in_block,
|
|
+ int *found_chunks,
|
|
+ u8 *chunk_data,
|
|
+ struct list_head *hard_list,
|
|
+ int summary_available)
|
|
+{
|
|
+ struct yaffs_obj_hdr *oh;
|
|
+ struct yaffs_obj *in;
|
|
+ struct yaffs_obj *parent;
|
|
+ int equiv_id;
|
|
+ loff_t file_size;
|
|
+ int is_shrink;
|
|
+ int is_unlinked;
|
|
+ struct yaffs_ext_tags tags;
|
|
+ int result;
|
|
+ int alloc_failed = 0;
|
|
+ int chunk = blk * dev->param.chunks_per_block + chunk_in_block;
|
|
+ struct yaffs_file_var *file_var;
|
|
+ struct yaffs_hardlink_var *hl_var;
|
|
+ struct yaffs_symlink_var *sl_var;
|
|
+
|
|
+ if (summary_available) {
|
|
+ result = yaffs_summary_fetch(dev, &tags, chunk_in_block);
|
|
+ tags.seq_number = bi->seq_number;
|
|
+ }
|
|
+
|
|
+ if (!summary_available || tags.obj_id == 0) {
|
|
+ result = yaffs_rd_chunk_tags_nand(dev, chunk, NULL, &tags);
|
|
+ dev->tags_used++;
|
|
+ } else {
|
|
+ dev->summary_used++;
|
|
+ }
|
|
+
|
|
+ if (result == YAFFS_FAIL)
|
|
+ yaffs_trace(YAFFS_TRACE_SCAN,
|
|
+ "Could not get tags for chunk %d\n", chunk);
|
|
+ /* Let's have a good look at this chunk... */
|
|
+
|
|
+ if (!tags.chunk_used) {
|
|
+ /* An unassigned chunk in the block.
|
|
+ * If there are used chunks after this one, then
|
|
+ * it is a chunk that was skipped due to failing
|
|
+ * the erased check. Just skip it so that it can
|
|
+ * be deleted.
|
|
+ * But, more typically, We get here when this is
|
|
+ * an unallocated chunk and his means that
|
|
+ * either the block is empty or this is the one
|
|
+ * being allocated from
|
|
+ */
|
|
+
|
|
+ if (*found_chunks) {
|
|
+ /* This is a chunk that was skipped due
|
|
+ * to failing the erased check */
|
|
+ } else if (chunk_in_block == 0) {
|
|
+ /* We're looking at the first chunk in
|
|
+ * the block so the block is unused */
|
|
+ bi->block_state = YAFFS_BLOCK_STATE_EMPTY;
|
|
+ dev->n_erased_blocks++;
|
|
+ } else {
|
|
+ if (bi->block_state == YAFFS_BLOCK_STATE_NEEDS_SCAN ||
|
|
+ bi->block_state == YAFFS_BLOCK_STATE_ALLOCATING) {
|
|
+ if (dev->seq_number == bi->seq_number) {
|
|
+ /* Allocating from this block*/
|
|
+ yaffs_trace(YAFFS_TRACE_SCAN,
|
|
+ " Allocating from %d %d",
|
|
+ blk, chunk_in_block);
|
|
+
|
|
+ bi->block_state =
|
|
+ YAFFS_BLOCK_STATE_ALLOCATING;
|
|
+ dev->alloc_block = blk;
|
|
+ dev->alloc_page = chunk_in_block;
|
|
+ dev->alloc_block_finder = blk;
|
|
+ } else {
|
|
+ /* This is a partially written block
|
|
+ * that is not the current
|
|
+ * allocation block.
|
|
+ */
|
|
+ yaffs_trace(YAFFS_TRACE_SCAN,
|
|
+ "Partially written block %d detected. gc will fix this.",
|
|
+ blk);
|
|
+ }
|
|
+ }
|
|
+ }
|
|
+
|
|
+ dev->n_free_chunks++;
|
|
+
|
|
+ } else if (tags.ecc_result ==
|
|
+ YAFFS_ECC_RESULT_UNFIXED) {
|
|
+ yaffs_trace(YAFFS_TRACE_SCAN,
|
|
+ " Unfixed ECC in chunk(%d:%d), chunk ignored",
|
|
+ blk, chunk_in_block);
|
|
+ dev->n_free_chunks++;
|
|
+ } else if (tags.obj_id > YAFFS_MAX_OBJECT_ID ||
|
|
+ tags.chunk_id > YAFFS_MAX_CHUNK_ID ||
|
|
+ tags.obj_id == YAFFS_OBJECTID_SUMMARY ||
|
|
+ (tags.chunk_id > 0 &&
|
|
+ tags.n_bytes > dev->data_bytes_per_chunk) ||
|
|
+ tags.seq_number != bi->seq_number) {
|
|
+ yaffs_trace(YAFFS_TRACE_SCAN,
|
|
+ "Chunk (%d:%d) with bad tags:obj = %d, chunk_id = %d, n_bytes = %d, ignored",
|
|
+ blk, chunk_in_block, tags.obj_id,
|
|
+ tags.chunk_id, tags.n_bytes);
|
|
+ dev->n_free_chunks++;
|
|
+ } else if (tags.chunk_id > 0) {
|
|
+ /* chunk_id > 0 so it is a data chunk... */
|
|
+ loff_t endpos;
|
|
+ loff_t chunk_base = (tags.chunk_id - 1) *
|
|
+ dev->data_bytes_per_chunk;
|
|
+
|
|
+ *found_chunks = 1;
|
|
+
|
|
+ yaffs_set_chunk_bit(dev, blk, chunk_in_block);
|
|
+ bi->pages_in_use++;
|
|
+
|
|
+ in = yaffs_find_or_create_by_number(dev,
|
|
+ tags.obj_id,
|
|
+ YAFFS_OBJECT_TYPE_FILE);
|
|
+ if (!in)
|
|
+ /* Out of memory */
|
|
+ alloc_failed = 1;
|
|
+
|
|
+ if (in &&
|
|
+ in->variant_type == YAFFS_OBJECT_TYPE_FILE &&
|
|
+ chunk_base < in->variant.file_variant.shrink_size) {
|
|
+ /* This has not been invalidated by
|
|
+ * a resize */
|
|
+ if (!yaffs_put_chunk_in_file(in, tags.chunk_id,
|
|
+ chunk, -1))
|
|
+ alloc_failed = 1;
|
|
+
|
|
+ /* File size is calculated by looking at
|
|
+ * the data chunks if we have not
|
|
+ * seen an object header yet.
|
|
+ * Stop this practice once we find an
|
|
+ * object header.
|
|
+ */
|
|
+ endpos = chunk_base + tags.n_bytes;
|
|
+
|
|
+ if (!in->valid &&
|
|
+ in->variant.file_variant.stored_size < endpos) {
|
|
+ in->variant.file_variant.
|
|
+ stored_size = endpos;
|
|
+ in->variant.file_variant.
|
|
+ file_size = endpos;
|
|
+ }
|
|
+ } else if (in) {
|
|
+ /* This chunk has been invalidated by a
|
|
+ * resize, or a past file deletion
|
|
+ * so delete the chunk*/
|
|
+ yaffs_chunk_del(dev, chunk, 1, __LINE__);
|
|
+ }
|
|
+ } else {
|
|
+ /* chunk_id == 0, so it is an ObjectHeader.
|
|
+ * Thus, we read in the object header and make
|
|
+ * the object
|
|
+ */
|
|
+ *found_chunks = 1;
|
|
+
|
|
+ yaffs_set_chunk_bit(dev, blk, chunk_in_block);
|
|
+ bi->pages_in_use++;
|
|
+
|
|
+ oh = NULL;
|
|
+ in = NULL;
|
|
+
|
|
+ if (tags.extra_available) {
|
|
+ in = yaffs_find_or_create_by_number(dev,
|
|
+ tags.obj_id,
|
|
+ tags.extra_obj_type);
|
|
+ if (!in)
|
|
+ alloc_failed = 1;
|
|
+ }
|
|
+
|
|
+ if (!in ||
|
|
+ (!in->valid && dev->param.disable_lazy_load) ||
|
|
+ tags.extra_shadows ||
|
|
+ (!in->valid && (tags.obj_id == YAFFS_OBJECTID_ROOT ||
|
|
+ tags.obj_id == YAFFS_OBJECTID_LOSTNFOUND))) {
|
|
+
|
|
+ /* If we don't have valid info then we
|
|
+ * need to read the chunk
|
|
+ * TODO In future we can probably defer
|
|
+ * reading the chunk and living with
|
|
+ * invalid data until needed.
|
|
+ */
|
|
+
|
|
+ result = yaffs_rd_chunk_tags_nand(dev,
|
|
+ chunk,
|
|
+ chunk_data,
|
|
+ NULL);
|
|
+
|
|
+ oh = (struct yaffs_obj_hdr *)chunk_data;
|
|
+
|
|
+ yaffs_do_endian_oh(dev, oh);
|
|
+
|
|
+ if (dev->param.inband_tags) {
|
|
+ /* Fix up the header if they got
|
|
+ * corrupted by inband tags */
|
|
+ oh->shadows_obj =
|
|
+ oh->inband_shadowed_obj_id;
|
|
+ oh->is_shrink =
|
|
+ oh->inband_is_shrink;
|
|
+ }
|
|
+
|
|
+ if (!in) {
|
|
+ in = yaffs_find_or_create_by_number(dev,
|
|
+ tags.obj_id, oh->type);
|
|
+ if (!in)
|
|
+ alloc_failed = 1;
|
|
+ }
|
|
+ }
|
|
+
|
|
+ if (!in) {
|
|
+ /* TODO Hoosterman we have a problem! */
|
|
+ yaffs_trace(YAFFS_TRACE_ERROR,
|
|
+ "yaffs tragedy: Could not make object for object %d at chunk %d during scan",
|
|
+ tags.obj_id, chunk);
|
|
+ return YAFFS_FAIL;
|
|
+ }
|
|
+
|
|
+ if (in->valid) {
|
|
+ /* We have already filled this one.
|
|
+ * We have a duplicate that will be
|
|
+ * discarded, but we first have to suck
|
|
+ * out resize info if it is a file.
|
|
+ */
|
|
+ if ((in->variant_type == YAFFS_OBJECT_TYPE_FILE) &&
|
|
+ ((oh && oh->type == YAFFS_OBJECT_TYPE_FILE) ||
|
|
+ (tags.extra_available &&
|
|
+ tags.extra_obj_type == YAFFS_OBJECT_TYPE_FILE)
|
|
+ )) {
|
|
+ loff_t this_size = (oh) ?
|
|
+ yaffs_oh_to_size(dev, oh, 0) :
|
|
+ tags.extra_file_size;
|
|
+ u32 parent_obj_id = (oh) ?
|
|
+ (u32)oh->parent_obj_id :
|
|
+ tags.extra_parent_id;
|
|
+
|
|
+ is_shrink = (oh) ?
|
|
+ oh->is_shrink :
|
|
+ tags.extra_is_shrink;
|
|
+
|
|
+ /* If it is deleted (unlinked
|
|
+ * at start also means deleted)
|
|
+ * we treat the file size as
|
|
+ * being zeroed at this point.
|
|
+ */
|
|
+ if (parent_obj_id == YAFFS_OBJECTID_DELETED ||
|
|
+ parent_obj_id == YAFFS_OBJECTID_UNLINKED) {
|
|
+ this_size = 0;
|
|
+ is_shrink = 1;
|
|
+ }
|
|
+
|
|
+ if (is_shrink &&
|
|
+ in->variant.file_variant.shrink_size >
|
|
+ this_size)
|
|
+ in->variant.file_variant.shrink_size =
|
|
+ this_size;
|
|
+
|
|
+ if (is_shrink)
|
|
+ bi->has_shrink_hdr = 1;
|
|
+ }
|
|
+ /* Use existing - destroy this one. */
|
|
+ yaffs_chunk_del(dev, chunk, 1, __LINE__);
|
|
+ }
|
|
+
|
|
+ if (!in->valid && in->variant_type !=
|
|
+ (oh ? oh->type : tags.extra_obj_type)) {
|
|
+ yaffs_trace(YAFFS_TRACE_ERROR,
|
|
+ "yaffs tragedy: Bad type, %d != %d, for object %d at chunk %d during scan",
|
|
+ oh ? oh->type : tags.extra_obj_type,
|
|
+ in->variant_type, tags.obj_id,
|
|
+ chunk);
|
|
+ in = yaffs_retype_obj(in, oh ? oh->type : tags.extra_obj_type);
|
|
+ }
|
|
+
|
|
+ if (!in->valid &&
|
|
+ (tags.obj_id == YAFFS_OBJECTID_ROOT ||
|
|
+ tags.obj_id == YAFFS_OBJECTID_LOSTNFOUND)) {
|
|
+ /* We only load some info, don't fiddle
|
|
+ * with directory structure */
|
|
+ in->valid = 1;
|
|
+
|
|
+ if (oh) {
|
|
+ in->yst_mode = oh->yst_mode;
|
|
+ yaffs_load_attribs(in, oh);
|
|
+ in->lazy_loaded = 0;
|
|
+ } else {
|
|
+ in->lazy_loaded = 1;
|
|
+ }
|
|
+ in->hdr_chunk = chunk;
|
|
+
|
|
+ } else if (!in->valid) {
|
|
+ /* we need to load this info */
|
|
+ in->valid = 1;
|
|
+ in->hdr_chunk = chunk;
|
|
+ if (oh) {
|
|
+ in->variant_type = oh->type;
|
|
+ in->yst_mode = oh->yst_mode;
|
|
+ yaffs_load_attribs(in, oh);
|
|
+
|
|
+ if (oh->shadows_obj > 0)
|
|
+ yaffs_handle_shadowed_obj(dev,
|
|
+ oh->shadows_obj, 1);
|
|
+
|
|
+ yaffs_set_obj_name_from_oh(in, oh);
|
|
+ parent = yaffs_find_or_create_by_number(dev,
|
|
+ oh->parent_obj_id,
|
|
+ YAFFS_OBJECT_TYPE_DIRECTORY);
|
|
+ file_size = yaffs_oh_to_size(dev, oh, 0);
|
|
+ is_shrink = oh->is_shrink;
|
|
+ equiv_id = oh->equiv_id;
|
|
+ } else {
|
|
+ in->variant_type = tags.extra_obj_type;
|
|
+ parent = yaffs_find_or_create_by_number(dev,
|
|
+ tags.extra_parent_id,
|
|
+ YAFFS_OBJECT_TYPE_DIRECTORY);
|
|
+ file_size = tags.extra_file_size;
|
|
+ is_shrink = tags.extra_is_shrink;
|
|
+ equiv_id = tags.extra_equiv_id;
|
|
+ in->lazy_loaded = 1;
|
|
+ }
|
|
+ in->dirty = 0;
|
|
+
|
|
+ if (!parent)
|
|
+ alloc_failed = 1;
|
|
+
|
|
+ /* directory stuff...
|
|
+ * hook up to parent
|
|
+ */
|
|
+
|
|
+ if (parent &&
|
|
+ parent->variant_type == YAFFS_OBJECT_TYPE_UNKNOWN) {
|
|
+ /* Set up as a directory */
|
|
+ parent->variant_type =
|
|
+ YAFFS_OBJECT_TYPE_DIRECTORY;
|
|
+ INIT_LIST_HEAD(&parent->
|
|
+ variant.dir_variant.children);
|
|
+ } else if (!parent ||
|
|
+ parent->variant_type !=
|
|
+ YAFFS_OBJECT_TYPE_DIRECTORY) {
|
|
+ /* Hoosterman, another problem....
|
|
+ * Trying to use a non-directory as a directory
|
|
+ */
|
|
+
|
|
+ yaffs_trace(YAFFS_TRACE_ERROR,
|
|
+ "yaffs tragedy: attempting to use non-directory as a directory in scan. Put in lost+found."
|
|
+ );
|
|
+ parent = dev->lost_n_found;
|
|
+ }
|
|
+ yaffs_add_obj_to_dir(parent, in);
|
|
+
|
|
+ is_unlinked = (parent == dev->del_dir) ||
|
|
+ (parent == dev->unlinked_dir);
|
|
+
|
|
+ if (is_shrink)
|
|
+ /* Mark the block */
|
|
+ bi->has_shrink_hdr = 1;
|
|
+
|
|
+ /* Note re hardlinks.
|
|
+ * Since we might scan a hardlink before its equivalent
|
|
+ * object is scanned we put them all in a list.
|
|
+ * After scanning is complete, we should have all the
|
|
+ * objects, so we run through this list and fix up all
|
|
+ * the chains.
|
|
+ */
|
|
+
|
|
+ switch (in->variant_type) {
|
|
+ case YAFFS_OBJECT_TYPE_UNKNOWN:
|
|
+ /* Todo got a problem */
|
|
+ break;
|
|
+ case YAFFS_OBJECT_TYPE_FILE:
|
|
+ file_var = &in->variant.file_variant;
|
|
+ if (file_var->stored_size < file_size) {
|
|
+ /* This covers the case where the file
|
|
+ * size is greater than the data held.
|
|
+ * This will happen if the file is
|
|
+ * resized to be larger than its
|
|
+ * current data extents.
|
|
+ */
|
|
+ file_var->file_size = file_size;
|
|
+ file_var->stored_size = file_size;
|
|
+ }
|
|
+
|
|
+ if (file_var->shrink_size > file_size)
|
|
+ file_var->shrink_size = file_size;
|
|
+
|
|
+ break;
|
|
+ case YAFFS_OBJECT_TYPE_HARDLINK:
|
|
+ hl_var = &in->variant.hardlink_variant;
|
|
+ if (!is_unlinked) {
|
|
+ hl_var->equiv_id = equiv_id;
|
|
+ list_add(&in->hard_links, hard_list);
|
|
+ }
|
|
+ break;
|
|
+ case YAFFS_OBJECT_TYPE_DIRECTORY:
|
|
+ /* Do nothing */
|
|
+ break;
|
|
+ case YAFFS_OBJECT_TYPE_SPECIAL:
|
|
+ /* Do nothing */
|
|
+ break;
|
|
+ case YAFFS_OBJECT_TYPE_SYMLINK:
|
|
+ sl_var = &in->variant.symlink_variant;
|
|
+ if (oh) {
|
|
+ sl_var->alias =
|
|
+ yaffs_clone_str(oh->alias);
|
|
+ if (!sl_var->alias)
|
|
+ alloc_failed = 1;
|
|
+ }
|
|
+ break;
|
|
+ }
|
|
+ }
|
|
+ }
|
|
+ return alloc_failed ? YAFFS_FAIL : YAFFS_OK;
|
|
+}
|
|
+
|
|
+int yaffs2_scan_backwards(struct yaffs_dev *dev)
|
|
+{
|
|
+ u32 blk;
|
|
+ int block_iter;
|
|
+ int start_iter;
|
|
+ int end_iter;
|
|
+ int n_to_scan = 0;
|
|
+ enum yaffs_block_state state;
|
|
+ int c;
|
|
+ LIST_HEAD(hard_list);
|
|
+ struct yaffs_block_info *bi;
|
|
+ u32 seq_number;
|
|
+ int n_blocks = dev->internal_end_block - dev->internal_start_block + 1;
|
|
+ u8 *chunk_data;
|
|
+ int found_chunks;
|
|
+ int alloc_failed = 0;
|
|
+ struct yaffs_block_index *block_index = NULL;
|
|
+ int alt_block_index = 0;
|
|
+ int summary_available;
|
|
+
|
|
+ yaffs_trace(YAFFS_TRACE_SCAN,
|
|
+ "yaffs2_scan_backwards starts intstartblk %d intendblk %d...",
|
|
+ dev->internal_start_block, dev->internal_end_block);
|
|
+
|
|
+ dev->seq_number = YAFFS_LOWEST_SEQUENCE_NUMBER;
|
|
+
|
|
+ block_index =
|
|
+ kmalloc(n_blocks * sizeof(struct yaffs_block_index), GFP_NOFS);
|
|
+
|
|
+ if (!block_index) {
|
|
+ block_index =
|
|
+ vmalloc(n_blocks * sizeof(struct yaffs_block_index));
|
|
+ alt_block_index = 1;
|
|
+ }
|
|
+
|
|
+ if (!block_index) {
|
|
+ yaffs_trace(YAFFS_TRACE_SCAN,
|
|
+ "yaffs2_scan_backwards() could not allocate block index!"
|
|
+ );
|
|
+ return YAFFS_FAIL;
|
|
+ }
|
|
+
|
|
+ dev->blocks_in_checkpt = 0;
|
|
+
|
|
+ chunk_data = yaffs_get_temp_buffer(dev);
|
|
+
|
|
+ /* Scan all the blocks to determine their state */
|
|
+ bi = dev->block_info;
|
|
+ for (blk = dev->internal_start_block; blk <= dev->internal_end_block;
|
|
+ blk++) {
|
|
+ yaffs_clear_chunk_bits(dev, blk);
|
|
+ bi->pages_in_use = 0;
|
|
+ bi->soft_del_pages = 0;
|
|
+
|
|
+ yaffs_query_init_block_state(dev, blk, &state, &seq_number);
|
|
+
|
|
+ bi->block_state = state;
|
|
+ bi->seq_number = seq_number;
|
|
+
|
|
+ if (bi->seq_number == YAFFS_SEQUENCE_CHECKPOINT_DATA)
|
|
+ bi->block_state = YAFFS_BLOCK_STATE_CHECKPOINT;
|
|
+ if (bi->seq_number == YAFFS_SEQUENCE_BAD_BLOCK)
|
|
+ bi->block_state = YAFFS_BLOCK_STATE_DEAD;
|
|
+
|
|
+ yaffs_trace(YAFFS_TRACE_SCAN_DEBUG,
|
|
+ "Block scanning block %d state %d seq %d",
|
|
+ blk, bi->block_state, seq_number);
|
|
+
|
|
+ if (bi->block_state == YAFFS_BLOCK_STATE_CHECKPOINT) {
|
|
+ dev->blocks_in_checkpt++;
|
|
+
|
|
+ } else if (bi->block_state == YAFFS_BLOCK_STATE_DEAD) {
|
|
+ yaffs_trace(YAFFS_TRACE_BAD_BLOCKS,
|
|
+ "block %d is bad", blk);
|
|
+ } else if (bi->block_state == YAFFS_BLOCK_STATE_EMPTY) {
|
|
+ yaffs_trace(YAFFS_TRACE_SCAN_DEBUG, "Block empty ");
|
|
+ dev->n_erased_blocks++;
|
|
+ dev->n_free_chunks += dev->param.chunks_per_block;
|
|
+ } else if (bi->block_state ==
|
|
+ YAFFS_BLOCK_STATE_NEEDS_SCAN) {
|
|
+ /* Determine the highest sequence number */
|
|
+ if (seq_number >= YAFFS_LOWEST_SEQUENCE_NUMBER &&
|
|
+ seq_number < YAFFS_HIGHEST_SEQUENCE_NUMBER) {
|
|
+ block_index[n_to_scan].seq = seq_number;
|
|
+ block_index[n_to_scan].block = blk;
|
|
+ n_to_scan++;
|
|
+ if (seq_number >= dev->seq_number)
|
|
+ dev->seq_number = seq_number;
|
|
+ } else {
|
|
+ /* TODO: Nasty sequence number! */
|
|
+ yaffs_trace(YAFFS_TRACE_SCAN,
|
|
+ "Block scanning block %d has bad sequence number %d",
|
|
+ blk, seq_number);
|
|
+ }
|
|
+ }
|
|
+ bi++;
|
|
+ }
|
|
+
|
|
+ yaffs_trace(YAFFS_TRACE_ALWAYS, "%d blocks to be sorted...", n_to_scan);
|
|
+
|
|
+ cond_resched();
|
|
+
|
|
+ /* Sort the blocks by sequence number */
|
|
+ sort(block_index, n_to_scan, sizeof(struct yaffs_block_index),
|
|
+ yaffs2_ybicmp, NULL);
|
|
+
|
|
+ cond_resched();
|
|
+
|
|
+ yaffs_trace(YAFFS_TRACE_SCAN, "...done");
|
|
+
|
|
+ /* Now scan the blocks looking at the data. */
|
|
+ start_iter = 0;
|
|
+ end_iter = n_to_scan - 1;
|
|
+ yaffs_trace(YAFFS_TRACE_SCAN_DEBUG, "%d blocks to scan", n_to_scan);
|
|
+
|
|
+ /* For each block.... backwards */
|
|
+ for (block_iter = end_iter;
|
|
+ !alloc_failed && block_iter >= start_iter;
|
|
+ block_iter--) {
|
|
+ /* Cooperative multitasking! This loop can run for so
|
|
+ long that watchdog timers expire. */
|
|
+ cond_resched();
|
|
+
|
|
+ /* get the block to scan in the correct order */
|
|
+ blk = block_index[block_iter].block;
|
|
+ bi = yaffs_get_block_info(dev, blk);
|
|
+
|
|
+ summary_available = yaffs_summary_read(dev, dev->sum_tags, blk);
|
|
+
|
|
+ /* For each chunk in each block that needs scanning.... */
|
|
+ found_chunks = 0;
|
|
+ if (summary_available)
|
|
+ c = dev->chunks_per_summary - 1;
|
|
+ else
|
|
+ c = dev->param.chunks_per_block - 1;
|
|
+
|
|
+ for (/* c is already initialised */;
|
|
+ !alloc_failed && c >= 0 &&
|
|
+ (bi->block_state == YAFFS_BLOCK_STATE_NEEDS_SCAN ||
|
|
+ bi->block_state == YAFFS_BLOCK_STATE_ALLOCATING);
|
|
+ c--) {
|
|
+ /* Scan backwards...
|
|
+ * Read the tags and decide what to do
|
|
+ */
|
|
+ if (yaffs2_scan_chunk(dev, bi, blk, c,
|
|
+ &found_chunks, chunk_data,
|
|
+ &hard_list, summary_available) ==
|
|
+ YAFFS_FAIL)
|
|
+ alloc_failed = 1;
|
|
+ }
|
|
+
|
|
+ if (bi->block_state == YAFFS_BLOCK_STATE_NEEDS_SCAN) {
|
|
+ /* If we got this far while scanning, then the block
|
|
+ * is fully allocated. */
|
|
+ bi->block_state = YAFFS_BLOCK_STATE_FULL;
|
|
+ }
|
|
+
|
|
+ /* Now let's see if it was dirty */
|
|
+ if (bi->pages_in_use == 0 &&
|
|
+ !bi->has_shrink_hdr &&
|
|
+ bi->block_state == YAFFS_BLOCK_STATE_FULL) {
|
|
+ yaffs_block_became_dirty(dev, blk);
|
|
+ }
|
|
+ }
|
|
+
|
|
+ yaffs_skip_rest_of_block(dev);
|
|
+
|
|
+ if (alt_block_index)
|
|
+ vfree(block_index);
|
|
+ else
|
|
+ kfree(block_index);
|
|
+
|
|
+ /* Ok, we've done all the scanning.
|
|
+ * Fix up the hard link chains.
|
|
+ * We have scanned all the objects, now it's time to add these
|
|
+ * hardlinks.
|
|
+ */
|
|
+ yaffs_link_fixup(dev, &hard_list);
|
|
+
|
|
+ yaffs_release_temp_buffer(dev, chunk_data);
|
|
+
|
|
+ if (alloc_failed)
|
|
+ return YAFFS_FAIL;
|
|
+
|
|
+ yaffs_trace(YAFFS_TRACE_SCAN, "yaffs2_scan_backwards ends");
|
|
+
|
|
+ return YAFFS_OK;
|
|
+}
|