mirror of
https://github.com/edk2-porting/linux-next.git
synced 2025-01-08 13:44:01 +08:00
a126483e82
The 1Gb Macronix chip can have a maximum of 20 bad blocks, while
the 2Gb version has twice as many blocks and therefore the maximum
number of bad blocks is 40.
The 4Gb GigaDevice GD5F4GQ4xA has twice as many blocks as its 2Gb
counterpart and therefore a maximum of 80 bad blocks.
Fixes: 377e517b5f
("mtd: nand: Add max_bad_eraseblocks_per_lun info to memorg")
Reported-by: Emil Lenngren <emil.lenngren@gmail.com>
Signed-off-by: Frieder Schrempf <frieder.schrempf@kontron.de>
Signed-off-by: Miquel Raynal <miquel.raynal@bootlin.com>
232 lines
5.7 KiB
C
232 lines
5.7 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Author:
|
|
* Chuanhong Guo <gch981213@gmail.com>
|
|
*/
|
|
|
|
#include <linux/device.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/mtd/spinand.h>
|
|
|
|
#define SPINAND_MFR_GIGADEVICE 0xC8
|
|
#define GD5FXGQ4XA_STATUS_ECC_1_7_BITFLIPS (1 << 4)
|
|
#define GD5FXGQ4XA_STATUS_ECC_8_BITFLIPS (3 << 4)
|
|
|
|
#define GD5FXGQ4UEXXG_REG_STATUS2 0xf0
|
|
|
|
static SPINAND_OP_VARIANTS(read_cache_variants,
|
|
SPINAND_PAGE_READ_FROM_CACHE_QUADIO_OP(0, 2, NULL, 0),
|
|
SPINAND_PAGE_READ_FROM_CACHE_X4_OP(0, 1, NULL, 0),
|
|
SPINAND_PAGE_READ_FROM_CACHE_DUALIO_OP(0, 1, NULL, 0),
|
|
SPINAND_PAGE_READ_FROM_CACHE_X2_OP(0, 1, NULL, 0),
|
|
SPINAND_PAGE_READ_FROM_CACHE_OP(true, 0, 1, NULL, 0),
|
|
SPINAND_PAGE_READ_FROM_CACHE_OP(false, 0, 1, NULL, 0));
|
|
|
|
static SPINAND_OP_VARIANTS(write_cache_variants,
|
|
SPINAND_PROG_LOAD_X4(true, 0, NULL, 0),
|
|
SPINAND_PROG_LOAD(true, 0, NULL, 0));
|
|
|
|
static SPINAND_OP_VARIANTS(update_cache_variants,
|
|
SPINAND_PROG_LOAD_X4(false, 0, NULL, 0),
|
|
SPINAND_PROG_LOAD(false, 0, NULL, 0));
|
|
|
|
static int gd5fxgq4xa_ooblayout_ecc(struct mtd_info *mtd, int section,
|
|
struct mtd_oob_region *region)
|
|
{
|
|
if (section > 3)
|
|
return -ERANGE;
|
|
|
|
region->offset = (16 * section) + 8;
|
|
region->length = 8;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int gd5fxgq4xa_ooblayout_free(struct mtd_info *mtd, int section,
|
|
struct mtd_oob_region *region)
|
|
{
|
|
if (section > 3)
|
|
return -ERANGE;
|
|
|
|
if (section) {
|
|
region->offset = 16 * section;
|
|
region->length = 8;
|
|
} else {
|
|
/* section 0 has one byte reserved for bad block mark */
|
|
region->offset = 1;
|
|
region->length = 7;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int gd5fxgq4xa_ecc_get_status(struct spinand_device *spinand,
|
|
u8 status)
|
|
{
|
|
switch (status & STATUS_ECC_MASK) {
|
|
case STATUS_ECC_NO_BITFLIPS:
|
|
return 0;
|
|
|
|
case GD5FXGQ4XA_STATUS_ECC_1_7_BITFLIPS:
|
|
/* 1-7 bits are flipped. return the maximum. */
|
|
return 7;
|
|
|
|
case GD5FXGQ4XA_STATUS_ECC_8_BITFLIPS:
|
|
return 8;
|
|
|
|
case STATUS_ECC_UNCOR_ERROR:
|
|
return -EBADMSG;
|
|
|
|
default:
|
|
break;
|
|
}
|
|
|
|
return -EINVAL;
|
|
}
|
|
|
|
static int gd5fxgq4uexxg_ooblayout_ecc(struct mtd_info *mtd, int section,
|
|
struct mtd_oob_region *region)
|
|
{
|
|
if (section)
|
|
return -ERANGE;
|
|
|
|
region->offset = 64;
|
|
region->length = 64;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int gd5fxgq4uexxg_ooblayout_free(struct mtd_info *mtd, int section,
|
|
struct mtd_oob_region *region)
|
|
{
|
|
if (section)
|
|
return -ERANGE;
|
|
|
|
/* Reserve 1 bytes for the BBM. */
|
|
region->offset = 1;
|
|
region->length = 63;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int gd5fxgq4uexxg_ecc_get_status(struct spinand_device *spinand,
|
|
u8 status)
|
|
{
|
|
u8 status2;
|
|
struct spi_mem_op op = SPINAND_GET_FEATURE_OP(GD5FXGQ4UEXXG_REG_STATUS2,
|
|
&status2);
|
|
int ret;
|
|
|
|
switch (status & STATUS_ECC_MASK) {
|
|
case STATUS_ECC_NO_BITFLIPS:
|
|
return 0;
|
|
|
|
case GD5FXGQ4XA_STATUS_ECC_1_7_BITFLIPS:
|
|
/*
|
|
* Read status2 register to determine a more fine grained
|
|
* bit error status
|
|
*/
|
|
ret = spi_mem_exec_op(spinand->spimem, &op);
|
|
if (ret)
|
|
return ret;
|
|
|
|
/*
|
|
* 4 ... 7 bits are flipped (1..4 can't be detected, so
|
|
* report the maximum of 4 in this case
|
|
*/
|
|
/* bits sorted this way (3...0): ECCS1,ECCS0,ECCSE1,ECCSE0 */
|
|
return ((status & STATUS_ECC_MASK) >> 2) |
|
|
((status2 & STATUS_ECC_MASK) >> 4);
|
|
|
|
case GD5FXGQ4XA_STATUS_ECC_8_BITFLIPS:
|
|
return 8;
|
|
|
|
case STATUS_ECC_UNCOR_ERROR:
|
|
return -EBADMSG;
|
|
|
|
default:
|
|
break;
|
|
}
|
|
|
|
return -EINVAL;
|
|
}
|
|
|
|
static const struct mtd_ooblayout_ops gd5fxgq4xa_ooblayout = {
|
|
.ecc = gd5fxgq4xa_ooblayout_ecc,
|
|
.free = gd5fxgq4xa_ooblayout_free,
|
|
};
|
|
|
|
static const struct mtd_ooblayout_ops gd5fxgq4uexxg_ooblayout = {
|
|
.ecc = gd5fxgq4uexxg_ooblayout_ecc,
|
|
.free = gd5fxgq4uexxg_ooblayout_free,
|
|
};
|
|
|
|
static const struct spinand_info gigadevice_spinand_table[] = {
|
|
SPINAND_INFO("GD5F1GQ4xA", 0xF1,
|
|
NAND_MEMORG(1, 2048, 64, 64, 1024, 20, 1, 1, 1),
|
|
NAND_ECCREQ(8, 512),
|
|
SPINAND_INFO_OP_VARIANTS(&read_cache_variants,
|
|
&write_cache_variants,
|
|
&update_cache_variants),
|
|
0,
|
|
SPINAND_ECCINFO(&gd5fxgq4xa_ooblayout,
|
|
gd5fxgq4xa_ecc_get_status)),
|
|
SPINAND_INFO("GD5F2GQ4xA", 0xF2,
|
|
NAND_MEMORG(1, 2048, 64, 64, 2048, 40, 1, 1, 1),
|
|
NAND_ECCREQ(8, 512),
|
|
SPINAND_INFO_OP_VARIANTS(&read_cache_variants,
|
|
&write_cache_variants,
|
|
&update_cache_variants),
|
|
0,
|
|
SPINAND_ECCINFO(&gd5fxgq4xa_ooblayout,
|
|
gd5fxgq4xa_ecc_get_status)),
|
|
SPINAND_INFO("GD5F4GQ4xA", 0xF4,
|
|
NAND_MEMORG(1, 2048, 64, 64, 4096, 80, 1, 1, 1),
|
|
NAND_ECCREQ(8, 512),
|
|
SPINAND_INFO_OP_VARIANTS(&read_cache_variants,
|
|
&write_cache_variants,
|
|
&update_cache_variants),
|
|
0,
|
|
SPINAND_ECCINFO(&gd5fxgq4xa_ooblayout,
|
|
gd5fxgq4xa_ecc_get_status)),
|
|
SPINAND_INFO("GD5F1GQ4UExxG", 0xd1,
|
|
NAND_MEMORG(1, 2048, 128, 64, 1024, 20, 1, 1, 1),
|
|
NAND_ECCREQ(8, 512),
|
|
SPINAND_INFO_OP_VARIANTS(&read_cache_variants,
|
|
&write_cache_variants,
|
|
&update_cache_variants),
|
|
0,
|
|
SPINAND_ECCINFO(&gd5fxgq4uexxg_ooblayout,
|
|
gd5fxgq4uexxg_ecc_get_status)),
|
|
};
|
|
|
|
static int gigadevice_spinand_detect(struct spinand_device *spinand)
|
|
{
|
|
u8 *id = spinand->id.data;
|
|
int ret;
|
|
|
|
/*
|
|
* For GD NANDs, There is an address byte needed to shift in before IDs
|
|
* are read out, so the first byte in raw_id is dummy.
|
|
*/
|
|
if (id[1] != SPINAND_MFR_GIGADEVICE)
|
|
return 0;
|
|
|
|
ret = spinand_match_and_init(spinand, gigadevice_spinand_table,
|
|
ARRAY_SIZE(gigadevice_spinand_table),
|
|
id[2]);
|
|
if (ret)
|
|
return ret;
|
|
|
|
return 1;
|
|
}
|
|
|
|
static const struct spinand_manufacturer_ops gigadevice_spinand_manuf_ops = {
|
|
.detect = gigadevice_spinand_detect,
|
|
};
|
|
|
|
const struct spinand_manufacturer gigadevice_spinand_manufacturer = {
|
|
.id = SPINAND_MFR_GIGADEVICE,
|
|
.name = "GigaDevice",
|
|
.ops = &gigadevice_spinand_manuf_ops,
|
|
};
|