Index: head/sys/dev/cfi/cfi_core.c =================================================================== --- head/sys/dev/cfi/cfi_core.c (revision 345463) +++ head/sys/dev/cfi/cfi_core.c (revision 345464) @@ -1,930 +1,1018 @@ /*- * SPDX-License-Identifier: BSD-3-Clause * * Copyright (c) 2007, Juniper Networks, Inc. * Copyright (c) 2012-2013, SRI International * All rights reserved. * * Portions of this software were developed by SRI International and the * University of Cambridge Computer Laboratory under DARPA/AFRL contract * (FA8750-10-C-0237) ("CTSRD"), as part of the DARPA CRASH research * programme. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * 3. Neither the name of the author nor the names of any co-contributors * may be used to endorse or promote products derived from this software * without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED * AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. */ #include __FBSDID("$FreeBSD$"); #include "opt_cfi.h" #include #include #include #include #include #include #include #include #include #include #include #include #include #include static void cfi_add_sysctls(struct cfi_softc *); extern struct cdevsw cfi_cdevsw; char cfi_driver_name[] = "cfi"; devclass_t cfi_devclass; devclass_t cfi_diskclass; uint32_t cfi_read_raw(struct cfi_softc *sc, u_int ofs) { uint32_t val; ofs &= ~(sc->sc_width - 1); switch (sc->sc_width) { case 1: val = bus_space_read_1(sc->sc_tag, sc->sc_handle, ofs); break; case 2: val = bus_space_read_2(sc->sc_tag, sc->sc_handle, ofs); break; case 4: val = bus_space_read_4(sc->sc_tag, sc->sc_handle, ofs); break; default: val = ~0; break; } return (val); } uint32_t cfi_read(struct cfi_softc *sc, u_int ofs) { uint32_t val; uint16_t sval; ofs &= ~(sc->sc_width - 1); switch (sc->sc_width) { case 1: val = bus_space_read_1(sc->sc_tag, sc->sc_handle, ofs); break; case 2: sval = bus_space_read_2(sc->sc_tag, sc->sc_handle, ofs); #ifdef CFI_HARDWAREBYTESWAP val = sval; #else val = le16toh(sval); #endif break; case 4: val = bus_space_read_4(sc->sc_tag, sc->sc_handle, ofs); #ifndef CFI_HARDWAREBYTESWAP val = le32toh(val); #endif break; default: val = ~0; break; } return (val); } static void cfi_write(struct cfi_softc *sc, u_int ofs, u_int val) { ofs &= ~(sc->sc_width - 1); switch (sc->sc_width) { case 1: bus_space_write_1(sc->sc_tag, sc->sc_handle, ofs, val); break; case 2: #ifdef CFI_HARDWAREBYTESWAP bus_space_write_2(sc->sc_tag, sc->sc_handle, ofs, val); #else bus_space_write_2(sc->sc_tag, sc->sc_handle, ofs, htole16(val)); #endif break; case 4: #ifdef CFI_HARDWAREBYTESWAP bus_space_write_4(sc->sc_tag, sc->sc_handle, ofs, val); #else bus_space_write_4(sc->sc_tag, sc->sc_handle, ofs, htole32(val)); #endif break; } } /* * This is same workaound as NetBSD sys/dev/nor/cfi.c cfi_reset_default() */ static void cfi_reset_default(struct cfi_softc *sc) { cfi_write(sc, 0, CFI_BCS_READ_ARRAY2); cfi_write(sc, 0, CFI_BCS_READ_ARRAY); } uint8_t cfi_read_qry(struct cfi_softc *sc, u_int ofs) { uint8_t val; cfi_write(sc, CFI_QRY_CMD_ADDR * sc->sc_width, CFI_QRY_CMD_DATA); val = cfi_read(sc, ofs * sc->sc_width); cfi_reset_default(sc); return (val); } static void cfi_amd_write(struct cfi_softc *sc, u_int ofs, u_int addr, u_int data) { cfi_write(sc, ofs + AMD_ADDR_START, CFI_AMD_UNLOCK); cfi_write(sc, ofs + AMD_ADDR_ACK, CFI_AMD_UNLOCK_ACK); cfi_write(sc, ofs + addr, data); } static char * cfi_fmtsize(uint32_t sz) { static char buf[8]; static const char *sfx[] = { "", "K", "M", "G" }; int sfxidx; sfxidx = 0; while (sfxidx < 3 && sz > 1023) { sz /= 1024; sfxidx++; } sprintf(buf, "%u%sB", sz, sfx[sfxidx]); return (buf); } int cfi_probe(device_t dev) { char desc[80]; struct cfi_softc *sc; char *vend_str; int error; uint16_t iface, vend; sc = device_get_softc(dev); sc->sc_dev = dev; sc->sc_rid = 0; sc->sc_res = bus_alloc_resource_any(dev, SYS_RES_MEMORY, &sc->sc_rid, RF_ACTIVE); if (sc->sc_res == NULL) return (ENXIO); sc->sc_tag = rman_get_bustag(sc->sc_res); sc->sc_handle = rman_get_bushandle(sc->sc_res); if (sc->sc_width == 0) { sc->sc_width = 1; while (sc->sc_width <= 4) { if (cfi_read_qry(sc, CFI_QRY_IDENT) == 'Q') break; sc->sc_width <<= 1; } } else if (cfi_read_qry(sc, CFI_QRY_IDENT) != 'Q') { error = ENXIO; goto out; } if (sc->sc_width > 4) { error = ENXIO; goto out; } /* We got a Q. Check if we also have the R and the Y. */ if (cfi_read_qry(sc, CFI_QRY_IDENT + 1) != 'R' || cfi_read_qry(sc, CFI_QRY_IDENT + 2) != 'Y') { error = ENXIO; goto out; } /* Get the vendor and command set. */ vend = cfi_read_qry(sc, CFI_QRY_VEND) | (cfi_read_qry(sc, CFI_QRY_VEND + 1) << 8); sc->sc_cmdset = vend; switch (vend) { case CFI_VEND_AMD_ECS: case CFI_VEND_AMD_SCS: vend_str = "AMD/Fujitsu"; break; case CFI_VEND_INTEL_ECS: vend_str = "Intel/Sharp"; break; case CFI_VEND_INTEL_SCS: vend_str = "Intel"; break; case CFI_VEND_MITSUBISHI_ECS: case CFI_VEND_MITSUBISHI_SCS: vend_str = "Mitsubishi"; break; default: vend_str = "Unknown vendor"; break; } /* Get the device size. */ sc->sc_size = 1U << cfi_read_qry(sc, CFI_QRY_SIZE); /* Sanity-check the I/F */ iface = cfi_read_qry(sc, CFI_QRY_IFACE) | (cfi_read_qry(sc, CFI_QRY_IFACE + 1) << 8); /* * Adding 1 to iface will give us a bit-wise "switch" * that allows us to test for the interface width by * testing a single bit. */ iface++; error = (iface & sc->sc_width) ? 0 : EINVAL; if (error) goto out; snprintf(desc, sizeof(desc), "%s - %s", vend_str, cfi_fmtsize(sc->sc_size)); device_set_desc_copy(dev, desc); out: bus_release_resource(dev, SYS_RES_MEMORY, sc->sc_rid, sc->sc_res); return (error); } int cfi_attach(device_t dev) { struct cfi_softc *sc; u_int blksz, blocks; u_int r, u; uint64_t mtoexp, ttoexp; #ifdef CFI_SUPPORT_STRATAFLASH uint64_t ppr; char name[KENV_MNAMELEN], value[32]; #endif sc = device_get_softc(dev); sc->sc_dev = dev; sc->sc_rid = 0; sc->sc_res = bus_alloc_resource_any(dev, SYS_RES_MEMORY, &sc->sc_rid, #ifndef ATSE_CFI_HACK RF_ACTIVE); #else RF_ACTIVE | RF_SHAREABLE); #endif if (sc->sc_res == NULL) return (ENXIO); sc->sc_tag = rman_get_bustag(sc->sc_res); sc->sc_handle = rman_get_bushandle(sc->sc_res); /* Get time-out values for erase, write, and buffer write. */ ttoexp = cfi_read_qry(sc, CFI_QRY_TTO_ERASE); mtoexp = cfi_read_qry(sc, CFI_QRY_MTO_ERASE); if (ttoexp == 0) { device_printf(dev, "erase timeout == 0, using 2^16ms\n"); ttoexp = 16; } if (ttoexp > 41) { device_printf(dev, "insane timeout: 2^%jdms\n", ttoexp); return (EINVAL); } if (mtoexp == 0) { device_printf(dev, "max erase timeout == 0, using 2^%jdms\n", ttoexp + 4); mtoexp = 4; } if (ttoexp + mtoexp > 41) { device_printf(dev, "insane max erase timeout: 2^%jd\n", ttoexp + mtoexp); return (EINVAL); } sc->sc_typical_timeouts[CFI_TIMEOUT_ERASE] = SBT_1MS * (1ULL << ttoexp); sc->sc_max_timeouts[CFI_TIMEOUT_ERASE] = sc->sc_typical_timeouts[CFI_TIMEOUT_ERASE] * (1ULL << mtoexp); ttoexp = cfi_read_qry(sc, CFI_QRY_TTO_WRITE); mtoexp = cfi_read_qry(sc, CFI_QRY_MTO_WRITE); if (ttoexp == 0) { device_printf(dev, "write timeout == 0, using 2^18ns\n"); ttoexp = 18; } if (ttoexp > 51) { device_printf(dev, "insane write timeout: 2^%jdus\n", ttoexp); return (EINVAL); } if (mtoexp == 0) { device_printf(dev, "max write timeout == 0, using 2^%jdms\n", ttoexp + 4); mtoexp = 4; } if (ttoexp + mtoexp > 51) { device_printf(dev, "insane max write timeout: 2^%jdus\n", ttoexp + mtoexp); return (EINVAL); } sc->sc_typical_timeouts[CFI_TIMEOUT_WRITE] = SBT_1US * (1ULL << ttoexp); sc->sc_max_timeouts[CFI_TIMEOUT_WRITE] = sc->sc_typical_timeouts[CFI_TIMEOUT_WRITE] * (1ULL << mtoexp); ttoexp = cfi_read_qry(sc, CFI_QRY_TTO_BUFWRITE); mtoexp = cfi_read_qry(sc, CFI_QRY_MTO_BUFWRITE); /* Don't check for 0, it means not-supported. */ if (ttoexp > 51) { device_printf(dev, "insane write timeout: 2^%jdus\n", ttoexp); return (EINVAL); } if (ttoexp + mtoexp > 51) { device_printf(dev, "insane max write timeout: 2^%jdus\n", ttoexp + mtoexp); return (EINVAL); } sc->sc_typical_timeouts[CFI_TIMEOUT_BUFWRITE] = SBT_1US * (1ULL << cfi_read_qry(sc, CFI_QRY_TTO_BUFWRITE)); sc->sc_max_timeouts[CFI_TIMEOUT_BUFWRITE] = sc->sc_typical_timeouts[CFI_TIMEOUT_BUFWRITE] * (1ULL << cfi_read_qry(sc, CFI_QRY_MTO_BUFWRITE)); /* Get the maximum size of a multibyte program */ if (sc->sc_typical_timeouts[CFI_TIMEOUT_BUFWRITE] != 0) sc->sc_maxbuf = 1 << (cfi_read_qry(sc, CFI_QRY_MAXBUF) | cfi_read_qry(sc, CFI_QRY_MAXBUF) << 8); else sc->sc_maxbuf = 0; /* Get erase regions. */ sc->sc_regions = cfi_read_qry(sc, CFI_QRY_NREGIONS); sc->sc_region = malloc(sc->sc_regions * sizeof(struct cfi_region), M_TEMP, M_WAITOK | M_ZERO); for (r = 0; r < sc->sc_regions; r++) { blocks = cfi_read_qry(sc, CFI_QRY_REGION(r)) | (cfi_read_qry(sc, CFI_QRY_REGION(r) + 1) << 8); sc->sc_region[r].r_blocks = blocks + 1; blksz = cfi_read_qry(sc, CFI_QRY_REGION(r) + 2) | (cfi_read_qry(sc, CFI_QRY_REGION(r) + 3) << 8); sc->sc_region[r].r_blksz = (blksz == 0) ? 128 : blksz * 256; } /* Reset the device to a default state. */ cfi_write(sc, 0, CFI_BCS_CLEAR_STATUS); if (bootverbose) { device_printf(dev, "["); for (r = 0; r < sc->sc_regions; r++) { printf("%ux%s%s", sc->sc_region[r].r_blocks, cfi_fmtsize(sc->sc_region[r].r_blksz), (r == sc->sc_regions - 1) ? "]\n" : ","); } } + if (sc->sc_cmdset == CFI_VEND_AMD_ECS || + sc->sc_cmdset == CFI_VEND_AMD_SCS) { + cfi_amd_write(sc, 0, AMD_ADDR_START, CFI_AMD_AUTO_SELECT); + sc->sc_manid = cfi_read(sc, 0); + sc->sc_devid = cfi_read(sc, 2); + device_printf(dev, "Manufacturer ID:%x Device ID:%x\n", + sc->sc_manid, sc->sc_devid); + cfi_write(sc, 0, CFI_BCS_READ_ARRAY2); + } + u = device_get_unit(dev); sc->sc_nod = make_dev(&cfi_cdevsw, u, UID_ROOT, GID_WHEEL, 0600, "%s%u", cfi_driver_name, u); sc->sc_nod->si_drv1 = sc; cfi_add_sysctls(sc); #ifdef CFI_SUPPORT_STRATAFLASH /* * Store the Intel factory PPR in the environment. In some * cases it is the most unique ID on a board. */ if (cfi_intel_get_factory_pr(sc, &ppr) == 0) { if (snprintf(name, sizeof(name), "%s.factory_ppr", device_get_nameunit(dev)) < (sizeof(name) - 1) && snprintf(value, sizeof(value), "0x%016jx", ppr) < (sizeof(value) - 1)) (void) kern_setenv(name, value); } #endif device_add_child(dev, "cfid", -1); bus_generic_attach(dev); return (0); } static void cfi_add_sysctls(struct cfi_softc *sc) { struct sysctl_ctx_list *ctx; struct sysctl_oid_list *children; ctx = device_get_sysctl_ctx(sc->sc_dev); children = SYSCTL_CHILDREN(device_get_sysctl_tree(sc->sc_dev)); SYSCTL_ADD_UINT(ctx, children, OID_AUTO, "typical_erase_timout_count", CTLFLAG_RD, &sc->sc_tto_counts[CFI_TIMEOUT_ERASE], 0, "Number of times the typical erase timeout was exceeded"); SYSCTL_ADD_UINT(ctx, children, OID_AUTO, "max_erase_timout_count", CTLFLAG_RD, &sc->sc_mto_counts[CFI_TIMEOUT_ERASE], 0, "Number of times the maximum erase timeout was exceeded"); SYSCTL_ADD_UINT(ctx, children, OID_AUTO, "typical_write_timout_count", CTLFLAG_RD, &sc->sc_tto_counts[CFI_TIMEOUT_WRITE], 0, "Number of times the typical write timeout was exceeded"); SYSCTL_ADD_UINT(ctx, children, OID_AUTO, "max_write_timout_count", CTLFLAG_RD, &sc->sc_mto_counts[CFI_TIMEOUT_WRITE], 0, "Number of times the maximum write timeout was exceeded"); if (sc->sc_maxbuf > 0) { SYSCTL_ADD_UINT(ctx, children, OID_AUTO, "typical_bufwrite_timout_count", CTLFLAG_RD, &sc->sc_tto_counts[CFI_TIMEOUT_BUFWRITE], 0, "Number of times the typical buffered write timeout was " "exceeded"); SYSCTL_ADD_UINT(ctx, children, OID_AUTO, "max_bufwrite_timout_count", CTLFLAG_RD, &sc->sc_mto_counts[CFI_TIMEOUT_BUFWRITE], 0, "Number of times the maximum buffered write timeout was " "exceeded"); } } int cfi_detach(device_t dev) { struct cfi_softc *sc; sc = device_get_softc(dev); destroy_dev(sc->sc_nod); free(sc->sc_region, M_TEMP); bus_release_resource(dev, SYS_RES_MEMORY, sc->sc_rid, sc->sc_res); return (0); } +static bool +cfi_check_erase(struct cfi_softc *sc, u_int ofs, u_int sz) +{ + bool result; + int i; + uint32_t val; + + result = FALSE; + for (i = 0; i < sz; i += sc->sc_width) { + val = cfi_read(sc, ofs + i); + switch (sc->sc_width) { + case 1: + if (val != 0xff) + goto out; + continue; + case 2: + if (val != 0xffff) + goto out; + continue; + case 4: + if (val != 0xffffffff) + goto out; + continue; + } + } + result = TRUE; + +out: + return (result); +} + static int cfi_wait_ready(struct cfi_softc *sc, u_int ofs, sbintime_t start, enum cfi_wait_cmd cmd) { int done, error, tto_exceeded; uint32_t st0 = 0, st = 0; sbintime_t now; done = 0; error = 0; tto_exceeded = 0; while (!done && !error) { /* * Save time before we start so we always do one check * after the timeout has expired. */ now = sbinuptime(); switch (sc->sc_cmdset) { case CFI_VEND_INTEL_ECS: case CFI_VEND_INTEL_SCS: st = cfi_read(sc, ofs); done = (st & CFI_INTEL_STATUS_WSMS); if (done) { /* NB: bit 0 is reserved */ st &= ~(CFI_INTEL_XSTATUS_RSVD | CFI_INTEL_STATUS_WSMS | CFI_INTEL_STATUS_RSVD); if (st & CFI_INTEL_STATUS_DPS) error = EPERM; else if (st & CFI_INTEL_STATUS_PSLBS) error = EIO; else if (st & CFI_INTEL_STATUS_ECLBS) error = ENXIO; else if (st) error = EACCES; } break; case CFI_VEND_AMD_SCS: case CFI_VEND_AMD_ECS: st0 = cfi_read(sc, ofs); st = cfi_read(sc, ofs); done = ((st & 0x40) == (st0 & 0x40)) ? 1 : 0; break; } if (tto_exceeded || now > start + sc->sc_typical_timeouts[cmd]) { if (!tto_exceeded) { tto_exceeded = 1; sc->sc_tto_counts[cmd]++; #ifdef CFI_DEBUG_TIMEOUT device_printf(sc->sc_dev, "typical timeout exceeded (cmd %d)", cmd); #endif } if (now > start + sc->sc_max_timeouts[cmd]) { sc->sc_mto_counts[cmd]++; #ifdef CFI_DEBUG_TIMEOUT device_printf(sc->sc_dev, "max timeout exceeded (cmd %d)", cmd); #endif } } } if (!done && !error) error = ETIMEDOUT; if (error) printf("\nerror=%d (st 0x%x st0 0x%x)\n", error, st, st0); return (error); } int cfi_write_block(struct cfi_softc *sc) { union { uint8_t *x8; uint16_t *x16; uint32_t *x32; } ptr, cpyprt; register_t intr; - int error, i, neederase = 0; + int error, i, j, neederase = 0; uint32_t st; u_int wlen; sbintime_t start; + u_int minsz; + uint32_t val; /* Intel flash must be unlocked before modification */ switch (sc->sc_cmdset) { case CFI_VEND_INTEL_ECS: case CFI_VEND_INTEL_SCS: cfi_write(sc, sc->sc_wrofs, CFI_INTEL_LBS); cfi_write(sc, sc->sc_wrofs, CFI_INTEL_UB); cfi_write(sc, sc->sc_wrofs, CFI_BCS_READ_ARRAY); break; } /* Check if an erase is required. */ for (i = 0; i < sc->sc_wrbufsz; i++) if ((sc->sc_wrbuf[i] & sc->sc_wrbufcpy[i]) != sc->sc_wrbuf[i]) { neederase = 1; break; } if (neederase) { intr = intr_disable(); start = sbinuptime(); /* Erase the block. */ switch (sc->sc_cmdset) { case CFI_VEND_INTEL_ECS: case CFI_VEND_INTEL_SCS: cfi_write(sc, sc->sc_wrofs, CFI_BCS_BLOCK_ERASE); cfi_write(sc, sc->sc_wrofs, CFI_BCS_CONFIRM); break; case CFI_VEND_AMD_SCS: case CFI_VEND_AMD_ECS: + /* find minimum sector size */ + minsz = sc->sc_region[0].r_blksz; + for (i = 1; i < sc->sc_regions; i++) { + if (sc->sc_region[i].r_blksz < minsz) + minsz = sc->sc_region[i].r_blksz; + } cfi_amd_write(sc, sc->sc_wrofs, AMD_ADDR_START, CFI_AMD_ERASE_SECTOR); - cfi_amd_write(sc, sc->sc_wrofs, 0, CFI_AMD_BLOCK_ERASE); + cfi_amd_write(sc, sc->sc_wrofs, + sc->sc_wrofs >> (ffs(minsz) - 1), + CFI_AMD_BLOCK_ERASE); + for (i = 0; i < CFI_AMD_MAXCHK; ++i) { + if (cfi_check_erase(sc, sc->sc_wrofs, + sc->sc_wrbufsz)) + break; + DELAY(10); + } + if (i == CFI_AMD_MAXCHK) { + printf("\nCFI Sector Erase time out error\n"); + return (ENODEV); + } break; default: /* Better safe than sorry... */ intr_restore(intr); return (ENODEV); } intr_restore(intr); error = cfi_wait_ready(sc, sc->sc_wrofs, start, CFI_TIMEOUT_ERASE); if (error) goto out; } else error = 0; /* Write the block using a multibyte write if supported. */ ptr.x8 = sc->sc_wrbuf; cpyprt.x8 = sc->sc_wrbufcpy; if (sc->sc_maxbuf > sc->sc_width) { switch (sc->sc_cmdset) { case CFI_VEND_INTEL_ECS: case CFI_VEND_INTEL_SCS: for (i = 0; i < sc->sc_wrbufsz; i += wlen) { wlen = MIN(sc->sc_maxbuf, sc->sc_wrbufsz - i); intr = intr_disable(); start = sbinuptime(); do { cfi_write(sc, sc->sc_wrofs + i, CFI_BCS_BUF_PROG_SETUP); if (sbinuptime() > start + sc->sc_max_timeouts[CFI_TIMEOUT_BUFWRITE]) { error = ETIMEDOUT; goto out; } st = cfi_read(sc, sc->sc_wrofs + i); } while (! (st & CFI_INTEL_STATUS_WSMS)); cfi_write(sc, sc->sc_wrofs + i, (wlen / sc->sc_width) - 1); switch (sc->sc_width) { case 1: bus_space_write_region_1(sc->sc_tag, sc->sc_handle, sc->sc_wrofs + i, ptr.x8 + i, wlen); break; case 2: bus_space_write_region_2(sc->sc_tag, sc->sc_handle, sc->sc_wrofs + i, ptr.x16 + i / 2, wlen / 2); break; case 4: bus_space_write_region_4(sc->sc_tag, sc->sc_handle, sc->sc_wrofs + i, ptr.x32 + i / 4, wlen / 4); break; } cfi_write(sc, sc->sc_wrofs + i, CFI_BCS_CONFIRM); intr_restore(intr); error = cfi_wait_ready(sc, sc->sc_wrofs + i, start, CFI_TIMEOUT_BUFWRITE); if (error != 0) goto out; } goto out; default: /* Fall through to single word case */ break; } } /* Write the block one byte/word at a time. */ for (i = 0; i < sc->sc_wrbufsz; i += sc->sc_width) { /* Avoid writing unless we are actually changing bits */ if (!neederase) { switch (sc->sc_width) { case 1: if(*(ptr.x8 + i) == *(cpyprt.x8 + i)) continue; break; case 2: if(*(ptr.x16 + i / 2) == *(cpyprt.x16 + i / 2)) continue; break; case 4: if(*(ptr.x32 + i / 4) == *(cpyprt.x32 + i / 4)) continue; break; } } /* * Make sure the command to start a write and the * actual write happens back-to-back without any * excessive delays. */ intr = intr_disable(); start = sbinuptime(); switch (sc->sc_cmdset) { case CFI_VEND_INTEL_ECS: case CFI_VEND_INTEL_SCS: cfi_write(sc, sc->sc_wrofs + i, CFI_BCS_PROGRAM); break; case CFI_VEND_AMD_SCS: case CFI_VEND_AMD_ECS: cfi_amd_write(sc, 0, AMD_ADDR_START, CFI_AMD_PROGRAM); break; } switch (sc->sc_width) { case 1: bus_space_write_1(sc->sc_tag, sc->sc_handle, sc->sc_wrofs + i, *(ptr.x8 + i)); break; case 2: bus_space_write_2(sc->sc_tag, sc->sc_handle, sc->sc_wrofs + i, *(ptr.x16 + i / 2)); break; case 4: bus_space_write_4(sc->sc_tag, sc->sc_handle, sc->sc_wrofs + i, *(ptr.x32 + i / 4)); break; } intr_restore(intr); - error = cfi_wait_ready(sc, sc->sc_wrofs, start, - CFI_TIMEOUT_WRITE); - if (error) - goto out; + if (sc->sc_cmdset == CFI_VEND_AMD_ECS || + sc->sc_cmdset == CFI_VEND_AMD_SCS) { + for (j = 0; j < CFI_AMD_MAXCHK; ++j) { + switch (sc->sc_width) { + case 1: + val = *(ptr.x8 + i); + break; + case 2: + val = *(ptr.x16 + i / 2); + break; + case 4: + val = *(ptr.x32 + i / 4); + break; + } + + if (cfi_read(sc, sc->sc_wrofs + i) == val) + break; + + DELAY(10); + } + if (j == CFI_AMD_MAXCHK) { + printf("\nCFI Program Verify time out error\n"); + error = ENXIO; + goto out; + } + } else { + error = cfi_wait_ready(sc, sc->sc_wrofs, start, + CFI_TIMEOUT_WRITE); + if (error) + goto out; + } } /* error is 0. */ out: cfi_reset_default(sc); /* Relock Intel flash */ switch (sc->sc_cmdset) { case CFI_VEND_INTEL_ECS: case CFI_VEND_INTEL_SCS: cfi_write(sc, sc->sc_wrofs, CFI_INTEL_LBS); cfi_write(sc, sc->sc_wrofs, CFI_INTEL_LB); cfi_write(sc, sc->sc_wrofs, CFI_BCS_READ_ARRAY); break; } return (error); } #ifdef CFI_SUPPORT_STRATAFLASH /* * Intel StrataFlash Protection Register Support. * * The memory includes a 128-bit Protection Register that can be * used for security. There are two 64-bit segments; one is programmed * at the factory with a unique 64-bit number which is immutable. * The other segment is left blank for User (OEM) programming. * The User/OEM segment is One Time Programmable (OTP). It can also * be locked to prevent any further writes by setting bit 0 of the * Protection Lock Register (PLR). The PLR can written only once. */ static uint16_t cfi_get16(struct cfi_softc *sc, int off) { uint16_t v = bus_space_read_2(sc->sc_tag, sc->sc_handle, off<<1); return v; } #ifdef CFI_ARMEDANDDANGEROUS static void cfi_put16(struct cfi_softc *sc, int off, uint16_t v) { bus_space_write_2(sc->sc_tag, sc->sc_handle, off<<1, v); } #endif /* * Read the factory-defined 64-bit segment of the PR. */ int cfi_intel_get_factory_pr(struct cfi_softc *sc, uint64_t *id) { if (sc->sc_cmdset != CFI_VEND_INTEL_ECS) return EOPNOTSUPP; KASSERT(sc->sc_width == 2, ("sc_width %d", sc->sc_width)); cfi_write(sc, 0, CFI_INTEL_READ_ID); *id = ((uint64_t)cfi_get16(sc, CFI_INTEL_PR(0)))<<48 | ((uint64_t)cfi_get16(sc, CFI_INTEL_PR(1)))<<32 | ((uint64_t)cfi_get16(sc, CFI_INTEL_PR(2)))<<16 | ((uint64_t)cfi_get16(sc, CFI_INTEL_PR(3))); cfi_write(sc, 0, CFI_BCS_READ_ARRAY); return 0; } /* * Read the User/OEM 64-bit segment of the PR. */ int cfi_intel_get_oem_pr(struct cfi_softc *sc, uint64_t *id) { if (sc->sc_cmdset != CFI_VEND_INTEL_ECS) return EOPNOTSUPP; KASSERT(sc->sc_width == 2, ("sc_width %d", sc->sc_width)); cfi_write(sc, 0, CFI_INTEL_READ_ID); *id = ((uint64_t)cfi_get16(sc, CFI_INTEL_PR(4)))<<48 | ((uint64_t)cfi_get16(sc, CFI_INTEL_PR(5)))<<32 | ((uint64_t)cfi_get16(sc, CFI_INTEL_PR(6)))<<16 | ((uint64_t)cfi_get16(sc, CFI_INTEL_PR(7))); cfi_write(sc, 0, CFI_BCS_READ_ARRAY); return 0; } /* * Write the User/OEM 64-bit segment of the PR. * XXX should allow writing individual words/bytes */ int cfi_intel_set_oem_pr(struct cfi_softc *sc, uint64_t id) { #ifdef CFI_ARMEDANDDANGEROUS register_t intr; int i, error; sbintime_t start; #endif if (sc->sc_cmdset != CFI_VEND_INTEL_ECS) return EOPNOTSUPP; KASSERT(sc->sc_width == 2, ("sc_width %d", sc->sc_width)); #ifdef CFI_ARMEDANDDANGEROUS for (i = 7; i >= 4; i--, id >>= 16) { intr = intr_disable(); start = sbinuptime(); cfi_write(sc, 0, CFI_INTEL_PP_SETUP); cfi_put16(sc, CFI_INTEL_PR(i), id&0xffff); intr_restore(intr); error = cfi_wait_ready(sc, CFI_BCS_READ_STATUS, start, CFI_TIMEOUT_WRITE); if (error) break; } cfi_write(sc, 0, CFI_BCS_READ_ARRAY); return error; #else device_printf(sc->sc_dev, "%s: OEM PR not set, " "CFI_ARMEDANDDANGEROUS not configured\n", __func__); return ENXIO; #endif } /* * Read the contents of the Protection Lock Register. */ int cfi_intel_get_plr(struct cfi_softc *sc, uint32_t *plr) { if (sc->sc_cmdset != CFI_VEND_INTEL_ECS) return EOPNOTSUPP; KASSERT(sc->sc_width == 2, ("sc_width %d", sc->sc_width)); cfi_write(sc, 0, CFI_INTEL_READ_ID); *plr = cfi_get16(sc, CFI_INTEL_PLR); cfi_write(sc, 0, CFI_BCS_READ_ARRAY); return 0; } /* * Write the Protection Lock Register to lock down the * user-settable segment of the Protection Register. * NOTE: this operation is not reversible. */ int cfi_intel_set_plr(struct cfi_softc *sc) { #ifdef CFI_ARMEDANDDANGEROUS register_t intr; int error; sbintime_t start; #endif if (sc->sc_cmdset != CFI_VEND_INTEL_ECS) return EOPNOTSUPP; KASSERT(sc->sc_width == 2, ("sc_width %d", sc->sc_width)); #ifdef CFI_ARMEDANDDANGEROUS /* worthy of console msg */ device_printf(sc->sc_dev, "set PLR\n"); intr = intr_disable(); binuptime(&start); cfi_write(sc, 0, CFI_INTEL_PP_SETUP); cfi_put16(sc, CFI_INTEL_PLR, 0xFFFD); intr_restore(intr); error = cfi_wait_ready(sc, CFI_BCS_READ_STATUS, start, CFI_TIMEOUT_WRITE); cfi_write(sc, 0, CFI_BCS_READ_ARRAY); return error; #else device_printf(sc->sc_dev, "%s: PLR not set, " "CFI_ARMEDANDDANGEROUS not configured\n", __func__); return ENXIO; #endif } #endif /* CFI_SUPPORT_STRATAFLASH */ Index: head/sys/dev/cfi/cfi_reg.h =================================================================== --- head/sys/dev/cfi/cfi_reg.h (revision 345463) +++ head/sys/dev/cfi/cfi_reg.h (revision 345464) @@ -1,155 +1,158 @@ /*- * SPDX-License-Identifier: BSD-3-Clause * * Copyright (c) 2007, Juniper Networks, Inc. * Copyright (c) 2012-2013, SRI International * All rights reserved. * * Portions of this software were developed by SRI International and the * University of Cambridge Computer Laboratory under DARPA/AFRL contract * (FA8750-10-C-0237) ("CTSRD"), as part of the DARPA CRASH research * programme. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * 3. Neither the name of the author nor the names of any co-contributors * may be used to endorse or promote products derived from this software * without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED * AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. * * $FreeBSD$ */ #ifndef _DEV_CFI_REG_H_ #define _DEV_CFI_REG_H_ struct cfi_qry { u_char reserved[16]; u_char ident[3]; /* "QRY" */ u_char pri_vend[2]; u_char pri_vend_eqt[2]; u_char alt_vend[2]; u_char alt_vend_eqt[2]; /* System Interface Information. */ u_char min_vcc; u_char max_vcc; u_char min_vpp; u_char max_vpp; u_char tto_byte_write; /* 2**n microseconds. */ u_char tto_buf_write; /* 2**n microseconds. */ u_char tto_block_erase; /* 2**n milliseconds. */ u_char tto_chip_erase; /* 2**n milliseconds. */ u_char mto_byte_write; /* 2**n times typical t/o. */ u_char mto_buf_write; /* 2**n times typical t/o. */ u_char mto_block_erase; /* 2**n times typical t/o. */ u_char mto_chip_erase; /* 2**n times typical t/o. */ /* Device Geometry Definition. */ u_char size; /* 2**n bytes. */ u_char iface[2]; u_char max_buf_write_size[2]; /* 2**n. */ u_char nregions; /* Number of erase regions. */ u_char region[4]; /* Single entry. */ /* Additional entries follow. */ /* Primary Vendor-specific Extended Query table follows. */ /* Alternate Vendor-specific Extended Query table follows. */ }; #define CFI_QRY_CMD_ADDR 0x55 #define CFI_QRY_CMD_DATA 0x98 #define CFI_QRY_IDENT offsetof(struct cfi_qry, ident) #define CFI_QRY_VEND offsetof(struct cfi_qry, pri_vend) #define CFI_QRY_TTO_WRITE offsetof(struct cfi_qry, tto_byte_write) #define CFI_QRY_TTO_BUFWRITE offsetof(struct cfi_qry, tto_buf_write) #define CFI_QRY_TTO_ERASE offsetof(struct cfi_qry, tto_block_erase) #define CFI_QRY_MTO_WRITE offsetof(struct cfi_qry, mto_byte_write) #define CFI_QRY_MTO_BUFWRITE offsetof(struct cfi_qry, mto_buf_write) #define CFI_QRY_MTO_ERASE offsetof(struct cfi_qry, mto_block_erase) #define CFI_QRY_SIZE offsetof(struct cfi_qry, size) #define CFI_QRY_IFACE offsetof(struct cfi_qry, iface) #define CFI_QRY_MAXBUF offsetof(struct cfi_qry, max_buf_write_size) #define CFI_QRY_NREGIONS offsetof(struct cfi_qry, nregions) #define CFI_QRY_REGION0 offsetof(struct cfi_qry, region) #define CFI_QRY_REGION(x) (CFI_QRY_REGION0 + (x) * 4) #define CFI_VEND_NONE 0x0000 #define CFI_VEND_INTEL_ECS 0x0001 #define CFI_VEND_AMD_SCS 0x0002 #define CFI_VEND_INTEL_SCS 0x0003 #define CFI_VEND_AMD_ECS 0x0004 #define CFI_VEND_MITSUBISHI_SCS 0x0100 #define CFI_VEND_MITSUBISHI_ECS 0x0101 #define CFI_IFACE_X8 0x0000 #define CFI_IFACE_X16 0x0001 #define CFI_IFACE_X8X16 0x0002 #define CFI_IFACE_X32 0x0003 #define CFI_IFACE_X16X32 0x0005 /* Standard Command Set (aka Basic Command Set) */ #define CFI_BCS_BLOCK_ERASE 0x20 #define CFI_BCS_PROGRAM 0x40 #define CFI_BCS_CLEAR_STATUS 0x50 #define CFI_BCS_READ_STATUS 0x70 #define CFI_BCS_ERASE_SUSPEND 0xb0 #define CFI_BCS_ERASE_RESUME 0xd0 /* Equals CONFIRM */ #define CFI_BCS_CONFIRM 0xd0 #define CFI_BCS_BUF_PROG_SETUP 0xe8 #define CFI_BCS_READ_ARRAY 0xff #define CFI_BCS_READ_ARRAY2 0xf0 /* Intel commands. */ #define CFI_INTEL_LB 0x01 /* Lock Block */ #define CFI_INTEL_LBS 0x60 /* Lock Block Setup */ #define CFI_INTEL_READ_ID 0x90 /* Read Identifier */ #define CFI_INTEL_PP_SETUP 0xc0 /* Protection Program Setup */ #define CFI_INTEL_UB 0xd0 /* Unlock Block */ /* NB: these are addresses for 16-bit accesses */ #define CFI_INTEL_PLR 0x80 /* Protection Lock Register */ #define CFI_INTEL_PR(n) (0x81+(n)) /* Protection Register */ /* Status register definitions */ #define CFI_INTEL_STATUS_WSMS 0x0080 /* Write Machine Status */ #define CFI_INTEL_STATUS_ESS 0x0040 /* Erase Suspend Status */ #define CFI_INTEL_STATUS_ECLBS 0x0020 /* Erase and Clear Lock-Bit Status */ #define CFI_INTEL_STATUS_PSLBS 0x0010 /* Program and Set Lock-Bit Status */ #define CFI_INTEL_STATUS_VPENS 0x0008 /* Programming Voltage Status */ #define CFI_INTEL_STATUS_PSS 0x0004 /* Program Suspend Status */ #define CFI_INTEL_STATUS_DPS 0x0002 /* Device Protect Status */ #define CFI_INTEL_STATUS_RSVD 0x0001 /* reserved */ /* eXtended Status register definitions */ #define CFI_INTEL_XSTATUS_WBS 0x8000 /* Write Buffer Status */ #define CFI_INTEL_XSTATUS_RSVD 0x7f00 /* reserved */ /* AMD commands. */ #define CFI_AMD_BLOCK_ERASE 0x30 #define CFI_AMD_UNLOCK_ACK 0x55 #define CFI_AMD_ERASE_SECTOR 0x80 +#define CFI_AMD_AUTO_SELECT 0x90 #define CFI_AMD_PROGRAM 0xa0 #define CFI_AMD_UNLOCK 0xaa #define AMD_ADDR_START 0xaaa #define AMD_ADDR_ACK 0x555 + +#define CFI_AMD_MAXCHK 0x10000 #endif /* _DEV_CFI_REG_H_ */ Index: head/sys/dev/cfi/cfi_var.h =================================================================== --- head/sys/dev/cfi/cfi_var.h (revision 345463) +++ head/sys/dev/cfi/cfi_var.h (revision 345464) @@ -1,107 +1,110 @@ /*- * SPDX-License-Identifier: BSD-3-Clause * * Copyright (c) 2007, Juniper Networks, Inc. * Copyright (c) 2012-2013, SRI International * All rights reserved. * * Portions of this software were developed by SRI International and the * University of Cambridge Computer Laboratory under DARPA/AFRL contract * (FA8750-10-C-0237) ("CTSRD"), as part of the DARPA CRASH research * programme. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * 3. Neither the name of the author nor the names of any co-contributors * may be used to endorse or promote products derived from this software * without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED * AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. * * $FreeBSD$ */ #ifndef _DEV_CFI_VAR_H_ #define _DEV_CFI_VAR_H_ enum cfi_wait_cmd { CFI_TIMEOUT_ERASE, CFI_TIMEOUT_WRITE, CFI_TIMEOUT_BUFWRITE }; struct cfi_region { u_int r_blocks; u_int r_blksz; }; struct cfi_softc { device_t sc_dev; struct resource *sc_res; bus_space_handle_t sc_handle; bus_space_tag_t sc_tag; int sc_rid; u_int sc_size; /* Flash size. */ u_int sc_width; /* Interface width. */ u_int sc_regions; /* Erase regions. */ struct cfi_region *sc_region; /* Array of region info. */ u_int sc_cmdset; sbintime_t sc_typical_timeouts[3]; sbintime_t sc_max_timeouts[3]; u_int sc_tto_counts[3]; u_int sc_mto_counts[3]; u_int sc_maxbuf; struct cdev *sc_nod; struct proc *sc_opened; /* Process that has us opened. */ u_char *sc_wrbuf; u_char *sc_wrbufcpy; u_int sc_wrbufsz; u_int sc_wrofs; u_int sc_writing; + + u_int sc_manid; + u_int sc_devid; }; extern char cfi_driver_name[]; extern devclass_t cfi_devclass; extern devclass_t cfi_diskclass; int cfi_probe(device_t); int cfi_attach(device_t); int cfi_detach(device_t); uint32_t cfi_read_raw(struct cfi_softc *, u_int); uint32_t cfi_read(struct cfi_softc *, u_int); uint8_t cfi_read_qry(struct cfi_softc *, u_int); int cfi_write_block(struct cfi_softc *); int cfi_block_start(struct cfi_softc *, u_int); int cfi_block_finish(struct cfi_softc *); #ifdef CFI_SUPPORT_STRATAFLASH int cfi_intel_get_factory_pr(struct cfi_softc *sc, uint64_t *); int cfi_intel_get_oem_pr(struct cfi_softc *sc, uint64_t *); int cfi_intel_set_oem_pr(struct cfi_softc *sc, uint64_t); int cfi_intel_get_plr(struct cfi_softc *sc, uint32_t *); int cfi_intel_set_plr(struct cfi_softc *sc); #endif /* CFI_SUPPORT_STRATAFLASH */ #endif /* _DEV_CFI_VAR_H_ */