int nandmtd_ReadChunkFromNAND(struct yaffs_dev *dev, int chunkInNAND, u8 *data, struct yaffs_spare *spare) { struct mtd_info *mtd = (struct mtd_info *)(dev->driver_context); struct mtd_oob_ops ops; size_t dummy; int retval = 0; loff_t addr = ((loff_t) chunkInNAND) * dev->data_bytes_per_chunk; u8 spareAsBytes[8]; /* OOB */ if (data && !spare) retval = mtd->read(mtd, addr, dev->data_bytes_per_chunk, &dummy, data); else if (spare) { if (dev->param.use_nand_ecc) { ops.mode = MTD_OOB_AUTO; ops.ooblen = 8; /* temp hack */ } else { ops.mode = MTD_OOB_RAW; ops.ooblen = YAFFS_BYTES_PER_SPARE; } ops.len = data ? dev->data_bytes_per_chunk : ops.ooblen; ops.datbuf = data; ops.ooboffs = 0; ops.oobbuf = spareAsBytes; retval = mtd->read_oob(mtd, addr, &ops); if (dev->param.use_nand_ecc) translate_oob2spare(spare, spareAsBytes); } if (retval == 0) return YAFFS_OK; else return YAFFS_FAIL; }
int nandmtd_ReadChunkFromNAND(yaffs_Device * dev, int chunkInNAND, __u8 * data, yaffs_Spare * spare) { struct mtd_info *mtd = (struct mtd_info *)(dev->genericDevice); #if (LINUX_VERSION_CODE > KERNEL_VERSION(2,6,17)) struct mtd_oob_ops ops; #endif size_t dummy; int retval = 0; loff_t addr = ((loff_t) chunkInNAND) * dev->nDataBytesPerChunk; #if (LINUX_VERSION_CODE > KERNEL_VERSION(2,6,17)) __u8 spareAsBytes[8]; /* OOB */ if (data && !spare) retval = mtd->read(mtd, addr, dev->nDataBytesPerChunk, &dummy, data); else if (spare) { if (dev->useNANDECC) { ops.mode = MTD_OOB_AUTO; ops.ooblen = 8; /* temp hack */ } else { ops.mode = MTD_OOB_RAW; ops.ooblen = YAFFS_BYTES_PER_SPARE; } ops.len = data ? dev->nDataBytesPerChunk : ops.ooblen; ops.datbuf = data; ops.ooboffs = 0; ops.oobbuf = spareAsBytes; retval = mtd->read_oob(mtd, addr, &ops); if (dev->useNANDECC) translate_oob2spare(spare, spareAsBytes); } #else __u8 *spareAsBytes = (__u8 *) spare; if (data && spare) { if (dev->useNANDECC) { /* Careful, this call adds 2 ints */ /* to the end of the spare data. Calling function */ /* should allocate enough memory for spare, */ /* i.e. [YAFFS_BYTES_PER_SPARE+2*sizeof(int)]. */ retval = mtd->read_ecc(mtd, addr, dev->nDataBytesPerChunk, &dummy, data, spareAsBytes, &yaffs_oobinfo); } else { retval = mtd->read_ecc(mtd, addr, dev->nDataBytesPerChunk, &dummy, data, spareAsBytes, &yaffs_noeccinfo); } } else { if (data) retval = mtd->read(mtd, addr, dev->nDataBytesPerChunk, &dummy, data); if (spare) retval = mtd->read_oob(mtd, addr, YAFFS_BYTES_PER_SPARE, &dummy, spareAsBytes); } #endif if (retval == 0) return YAFFS_OK; else return YAFFS_FAIL; }
int nandmtd_ReadChunkFromNAND(yaffs_Device *dev, int chunkInNAND, __u8 *data, yaffs_Spare *spare) { struct mtd_info *mtd = (struct mtd_info *)(dev->genericDevice); #if (MTD_VERSION_CODE > MTD_VERSION(2, 6, 17)) struct mtd_oob_ops ops; #endif size_t dummy; int retval = 0; loff_t addr = ((loff_t) chunkInNAND) * dev->nDataBytesPerChunk; #if (MTD_VERSION_CODE > MTD_VERSION(2, 6, 17)) __u8 spareAsBytes[8]; if (data && !spare) retval = mtd->read(mtd, addr, dev->nDataBytesPerChunk, &dummy, data); else if (spare) { if (dev->useNANDECC) { ops.mode = MTD_OOB_AUTO; ops.ooblen = 8; } else { ops.mode = MTD_OOB_RAW; ops.ooblen = YAFFS_BYTES_PER_SPARE; } ops.len = data ? dev->nDataBytesPerChunk : ops.ooblen; ops.datbuf = data; ops.ooboffs = 0; ops.oobbuf = spareAsBytes; retval = mtd->read_oob(mtd, addr, &ops); if (dev->useNANDECC) translate_oob2spare(spare, spareAsBytes); } #else __u8 *spareAsBytes = (__u8 *) spare; if (data && spare) { if (dev->useNANDECC) { retval = mtd->read_ecc(mtd, addr, dev->nDataBytesPerChunk, &dummy, data, spareAsBytes, &yaffs_oobinfo); } else { retval = mtd->read_ecc(mtd, addr, dev->nDataBytesPerChunk, &dummy, data, spareAsBytes, &yaffs_noeccinfo); } } else { if (data) retval = mtd->read(mtd, addr, dev->nDataBytesPerChunk, &dummy, data); if (spare) retval = mtd->read_oob(mtd, addr, YAFFS_BYTES_PER_SPARE, &dummy, spareAsBytes); } #endif if (retval == 0) return YAFFS_OK; else return YAFFS_FAIL; }