WO1994023432A1 - Flashmemory mass storage architecture - Google Patents

Flashmemory mass storage architecture Download PDF

Info

Publication number
WO1994023432A1
WO1994023432A1 PCT/US1994/003154 US9403154W WO9423432A1 WO 1994023432 A1 WO1994023432 A1 WO 1994023432A1 US 9403154 W US9403154 W US 9403154W WO 9423432 A1 WO9423432 A1 WO 9423432A1
Authority
WO
WIPO (PCT)
Prior art keywords
block
count value
counter
erase
blocks
Prior art date
Application number
PCT/US1994/003154
Other languages
French (fr)
Inventor
Mahmud Assar
Siamack Nemazie
Petro Estakhri
Original Assignee
Cirrus Logic, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Family has litigation
First worldwide family litigation filed litigation Critical https://patents.darts-ip.com/?family=21901226&utm_source=google_patent&utm_medium=platform_link&utm_campaign=public_patent_search&patent=WO1994023432(A1) "Global patent litigation dataset” by Darts-ip is licensed under a Creative Commons Attribution 4.0 International License.
Application filed by Cirrus Logic, Inc. filed Critical Cirrus Logic, Inc.
Priority to EP94912849A priority Critical patent/EP0693216A4/en
Publication of WO1994023432A1 publication Critical patent/WO1994023432A1/en

Links

Classifications

    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11CSTATIC STORES
    • G11C16/00Erasable programmable read-only memories
    • G11C16/02Erasable programmable read-only memories electrically programmable
    • G11C16/06Auxiliary circuits, e.g. for writing into memory
    • G11C16/34Determination of programming status, e.g. threshold voltage, overprogramming or underprogramming, retention
    • G11C16/349Arrangements for evaluating degradation, retention or wearout, e.g. by counting erase cycles
    • G11C16/3495Circuits or methods to detect or delay wearout of nonvolatile EPROM or EEPROM memory devices, e.g. by counting numbers of erase or reprogram cycles, by using multiple memory areas serially or cyclically
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11CSTATIC STORES
    • G11C16/00Erasable programmable read-only memories
    • G11C16/02Erasable programmable read-only memories electrically programmable
    • G11C16/06Auxiliary circuits, e.g. for writing into memory
    • G11C16/10Programming or data input circuits
    • G11C16/102External programming circuits, e.g. EPROM programmers; In-circuit programming or reprogramming; EPROM emulators
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11CSTATIC STORES
    • G11C16/00Erasable programmable read-only memories
    • G11C16/02Erasable programmable read-only memories electrically programmable
    • G11C16/06Auxiliary circuits, e.g. for writing into memory
    • G11C16/34Determination of programming status, e.g. threshold voltage, overprogramming or underprogramming, retention
    • G11C16/349Arrangements for evaluating degradation, retention or wearout, e.g. by counting erase cycles
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0628Interfaces specially adapted for storage systems making use of a particular technique
    • G06F3/0662Virtualisation aspects
    • G06F3/0664Virtualisation aspects at device level, e.g. emulation of a storage device or system

Definitions

  • This invention relates to the field of mass storage for computers. More particularly, this invention relates to an architecture for replacing a hard disk with a semiconductor non-volatile memory and in particular flash memory.
  • Flash memory is a single transistor memory cell which is programmable through hot electron injection and erasable through Fowler-Nordheim tunneling. The programming and erasing of such a memory cell requires current to pass through the dielectric surrounding a floating gate electrode. Because of this, such types of memory have a finite number of erase-write cycles. Eventually, the dielectric will fail. Manufacturers of flash cell devices specify the limit for the number erase-write cycles as between 10,000 and 100,000. Accordingly, unlike rotating magnetic media, a flash memory mass storage device does not have an indefinite lifetime.
  • a semiconductor mass storage device Another requirement for a semiconductor mass storage device to be successful is that its use in lieu of a rotating media hard disk mass storage device be transparent to the system designer and the user. In other words, the designer of a computer incorporating such a semiconductor mass storage device could simply remove the hard disk and replace it with a semiconductor mass storage. All presently available commercial software should operate on a system employing such a semiconductor hard disk without the necessity of any modification.
  • SunDisk proposed an architecture for a semiconductor mass storage using flash memory at the Silicon Valley PC Design Conference July 9, 1991. That mass storage system included read-write block sizes of 512 Bytes (or multiples thereof) just like IBM PC compatible hard disk sector sizes. (IBM PC is a trademark of IBM Corporation.) During an erase cycle, an entire block is first fully programmed and then erased.
  • hard disk users typically store both information which is rarely changed and information which is frequently changed.
  • a commercial spread sheet or word processing software programs stored on a user's system are rarely, if ever, changed.
  • the spread sheet data files or word processing documents are frequently changed.
  • different sectors of a hard disk typically have dramatically different usage in terms of the number of times the information stored thereon is changed. While this disparity has no impact on a hard disk because of its insensitivity to data changes, in a flash memory device, this variance can cause sections of the mass storage to wear out and be unusable significantly sooner than other sections of the mass storage.
  • the present invention discloses two primary algorithms and an associated hardware architecture for a semiconductor mass storage device.
  • data file in this patent document refers to any computer file including commercial software, a user program, word processing software document, spread sheet file and the like.
  • the first algorithm provides means for avoiding an erase-before-write cycle when writing a modified data file back onto the mass storage device. Instead, no erase is performed and the modified data file is written onto an empty portion of the mass storage..
  • the second algorithm prevents any portion of the mass storage from being erased a substantially larger number of times than any other portion. This prevents any one block of the mass storage from failing and becoming unusable earlier than any other block thereby extending the life of the entire mass storage.
  • the semiconductor mass storage architecture has blocks sized to conform with commercial hard disk sector sizes.
  • the blocks are individually erasable.
  • the semiconductor mass storage of the present invention can be substituted for a rotating hard disk with no impact to the user, so that such a substitution will be transparent.
  • Means are provided for avoiding the erase-before-write cycle each time information stored in the mass storage is changed.
  • the erase cycle is understood to include, fully programming each bit in the block to be erased, and then erasing all the bits in the block.
  • the present invention includes a programmable map to maintain a correlation between the logical address 308 and the physical address 408 of the updated information files.
  • the mass storage will fill up because there have been no erase cycles. At such times, the mass storage needs to be cleaned up with a multi-sector erase as fully described in the detailed description below.
  • a counter tracks the number of times each block is erased.
  • a programmable maximum value for the counter is also provided. As the number of erase cycles for a block becomes one less than the maximum, the block is erased one last time and written with another file having a then smallest number of erase cycles. It is also prevented from being erased thereafter by setting its erase inhibit flag. After all blocks approach this maximum, all the erase counters and inhibit flags are cleared and the second algorithm is then repeated. In this way, no block can be erased a substantial number of times more than any other block.
  • flags are provided for defective blocks, used blocks, old version of a block, a count to determine the number of times a block has been erased and written and an erase inhibit flag.
  • Figure 1 shows an architecture for a semiconductor mass storage.
  • Figure 2 shows the architecture of Figure 1 wherein the data in one block has been altered and stored in a new physical address.
  • Figure 3 shows a block diagram of an erase cycle usage according to algorithm 1 of the present invention.
  • Figure 4 shows a simplified block diagram of the old/new flag system integrally formed with the memory.
  • Figure 5 shows a flow chart block diagram for algorithm 1.
  • Figure 6 shows an additional architecture according to the preferred embodiment of the present invention.
  • Figure 7 shows a flow chart block diagram of algorithm 2 of the present invention.
  • Figure 8 shows a flow chart block diagram of a read algorithm according to the present invention.
  • FIG. 1 shows an architecture for a semiconductor mass storage according to the present invention.
  • all of the memory storage is flash EEPROM. It is possible to substitute E 2 PROM for some or all of the data bits shown.
  • a memory storage 100 is arranged into N blocks of data from zero through N-l. Each of the blocks of data is M Bytes long. In the preferred embodiment, each block is 512 Bytes long to correspond with a sector length in a commercially available hard disk drive.
  • a flag 104 is directly related to each data block 102.
  • the memory 100 can contain as much memory storage as a user desires.
  • An example of a mass storage device might include 100 MByte of addressable storage.
  • a non-volatile content addressable memory (CAM) 106 is associated with the memory storage 100.
  • the CAM 106 is formed of flash memory.
  • the CAM 106 can also be E 2 PROM.
  • the CAM 106 is also formed of a non-volatile memory because loss of its information would make retrieval of the data files stored in the mass storage 100 impossible.
  • conventional computer systems are not configured to track continually changing physical locations of data files. According to the present invention, each time a data file is changed it is stored into a new physical location in the mass storage.
  • implementation of the architecture of the present invention requires a mapping of the logical address 308, i.e., the address where the computer system believes the data file is stored to the physical address 408, i.e., the actual location the data file can be found is stored in the mass storage.
  • the logical address 308 portion of the map 108 and the flags 112, 116 and 118 form part of the CAM 106. It is possible to use other storage means than a CAM to store the address map, such as a look-up table. However, a CAM is the most efficient means known to the inventors. It is not necessary that the physical address 408 portion of the map 108 form part of the CAM. Indeed, the physical address 408 portion of the- map 108 can be ordinary flash memory, E 2 PROM or even ROM. If ROM is selected for the physical address 408 array of the map 108, a defect in the ROM will prevent the block corresponding to that physical address 408 from ever being addressed. Accordingly, a changeable nonvolatile memory is preferred. Note that any replacement circuit for the CAM should be nonvolatile. Otherwise, loss or removal of power to the system will result in loss of the ability to find the data files in the mass storage.
  • the document will be stored in the mass storage system as shown in Figure 1.
  • the computer system will assign it a logical address 308, for example 526H.
  • the mass storage system of the present invention will select a physical address 408 of an unused block or blocks in the mass storage 100 for storing the document, e.g. 728H. That map correlating the logical address 308 to the physical address 408 is stored in the CAM 106.
  • the system of the present invention also sets the used/free flag 112 to indicate that this block has been written without being erased.
  • the used/free flag 112 also forms a portion of the CAM 106.
  • One used/free flag 112 is provided for each entry of the CAM 106.
  • the system of the present invention provides means for locating a block having its used/free flag 112 unset (not programmed) which indicates that the associated block is erased.
  • the system sets the used/free flag for the new block 114 ( Figure 2) and then stores the modified document in that new block 114.
  • the system sets the old/new flag 116 of the previous version of the document indicating that this is an old unneeded version of the document.
  • the system updates the correlation between the logical address 308 and the actual physical address 408.
  • the system of the present invention avoids the overhead of an erase cycle which is required in the erase-before-write of conventional systems to store a modified version of a previous document.
  • the writing to mass storage process outlined above is repeated until the entire mass storage memory 100 has been filled.
  • a full mass storage is indicated by no unset used/free flags 112 in the CAM 106.
  • a multi- sector erase is necessary and those blocks in the memory 100 and their associated CAM 106 entries having an old/new flag 116 set are all erased simultaneously. Note that it is not necessary for 100% of the blocks to have a set used/free flag 112 for a multi-sector erase to occur. For example, if a data file requiring three blocks were being written and only two blocks having unset used/free flags 112 were available a multi-sector erase can be run.
  • a simultaneous erase is not needed with prior art implementations because those embodiments utilize an erase- before-write cycle rather than retaining superseded versions of data files.
  • a latch of volatile logic circuits is set to couple the voltage necessary to erase the flash cells in the block. Because of the likely large number of memory blocks in the mass storage 100, if the CAM 106 and mass storage 100 are on the same integrated circuit (chip) coupling the old/new flag 116 to the latches in parallel would typically be very expensive in terms of surface area of the chip and coupling the old/new flags 116 serially to the latches would be expensive in terms of system performance.
  • a plurality of old/new flag systems 104 are intimately associated with each block in the memory 102 and is programmed by the same sequence of instructions as the old/new flag 116 of the CAM 106.
  • Figure 4 shows a simplified block diagram of the old/new flag system 104 which includes a non-volatile bit 120 having data which mirrors the old/new flag 116.
  • a volatile latch 122 coupled to receive the data in the bit 120 from the latch during an erase cycle.
  • the data in each of the bits 120 is simultaneously coupled to each appropriate ones of the latches 122 under control of a load signal coupled to each latch 122 over a load line L.
  • the latch for every block having its associated bit 120 set then couples the voltage necessary to perform an erase of that block and its associated bit 120.
  • all the latches 122 are individually reset to a predetermined state under control of a reset signal coupled to each latch 122 over a reset line R.
  • a simultaneous erase of all blocks having their respective old/new flags set may be undesirable.
  • the blocks can be segregated into groups of blocks. Each group has a unique control line to load the latches from the nonvolatile bits. In this mode, during an erase cycle, the control lines are sequentially activated and the groups of blocks sequentially erased.
  • Figure 5 shows algorithm 1 according to the present invention.
  • the system of the present invention receives an instruction to program data into the mass storage (step 200) , then the system attempts to locate a free block (step 202), i.e., a block having an unset (not programmed) used/free flag. If successful, the system sets the used/free flag for that block and programs the data into that block (step 206) .
  • a free block step 202
  • the system sets the used/free flag for that block and programs the data into that block (step 206) .
  • the system If on the other hand, the system is unable to locate a block having an unset used/free flag, the system erases the flags (used/free and old/new) and data for all blocks having a set old/new flag (step 204) and then searches for a block having an unset used/free flag (step 202) . Such a block has just been formed by step 204. The system then sets the used/free flag for that block and programs the data file into that block (step 206) . If the data file is a modified version of a previously existing file, the system must prevent the superseded version from being accessed. The.system determines whether the data file supersedes a previous data file (step 208) .
  • the system sets the old/new flag associated with the superseded block (step 210) . If on the other hand, the data file to be stored is a newly created data file, the step of setting the old/new flag (step 210) is skipped because there is no superseded block. Lastly, the map for correlating the logical address 308 to the physical address 408 is updated (step 212) .
  • the programming of the flash memory follows the procedure commonly understood by those of ordinary skill in the art.
  • the program impulses are appropriately applied to the bits to be programmed and then compared to the data being programmed to ensure that proper programming has occurred.
  • a defect flag 118 in the CAM 106 is set preventing that block from being used again.
  • FIG. 3 schematically shows the types of information stored in utilizing a mass storage media 150.
  • One portion of the mass storage 150 contains commercial applications software 152 such as word processing, spreadsheet, calendaring, calculators and the like. These portions of the mass storage 150 rarely, if ever, require an erase-reprogram cycle according to the algorithm described above.
  • a second section of the mass , storage 150 contains user data 154.
  • the user data 154 is frequently altered requiring the information to be reprogrammed into blocks of the free space 156 under the algorithm described above.
  • a third portion of the mass storage 150 contains free space 156 of unprogrammed blocks.
  • the storage blocks in the portions 154 and 156 of the memory 150 will recycle data files and thus be erased and reprogrammed significantly more often than the commercial applications software portion 152 of the memory 150. Accordingly, the mass storage 150 will wear out more quickly in the user data 154 and the free space 156 sections of the memory requiring earlier replacement than in sections 152 of the mass storage having data files which are rarely changed. As the number of free blocks diminishes providing a smaller number of blocks through which to recycle data files, the remaining blocks become erased more frequently exacerbating the problem.
  • a second algorithm is provided for leveling erase cycles amongst all the blocks within the entire mass storage device as shown in Figure 6.
  • a counter is provided for each block to count the number of times each block has been erased and reprogrammed.
  • An erase inhibit flag is also provided for each block. Once the erase count has reached the maximum for any block, the erase inhibit flag is set for that block. After that time that block cannot be erased until a clean-out erase is performed. Referring to Figure 3, if only algorithm 1 is used eventually all of the blocks in the user data 154 and the free space 156 portions of the mass storage 150 will reach the maximum count and have their respective erase inhibit flags set. Because of this, a reallocation of the rarely erased data files stored in the memory 152 is made into the memory 154 and/or 156.
  • a clean-out erase can be performed in the event that there is insufficient available storage for a data file presently being performed. For example, if all but two blocks have their respective erase inhibit flags set, and a three or more block data file is being programmed, a clean-out erase can be performed to provide sufficient storage for the data file.
  • the erase inhibit and erase count registers are erased and the cycle is repeated.
  • the selection of the maximum count depends upon the system requirements. As the value for the maximum count increases, the disparity between erase count cycles of various blocks can also increase. However, because data is shifted as a result of achieving maximum erase count this process of smoothing cycles throughout the mass storage of itself introduces additional erase cycles because a block of information is transferred from a physical block having few erases to a block having the maximum number of erases. Accordingly, though low maximum count values reduce the disparity between erase cycles amongst the blocks it also increases the number of erase cycles to which the blocks are subjected. Accordingly, individual users may select an erase count depending upon the system needs.
  • algorithm 2 is merged with algorithm 1 as shown in Figure 7.
  • An instruction is provided by the computer system to write a data file to the mass storage (step 230) which starts the combined algorithm 1 and algorithm 2 sequence. It is first determined whether the mass storage is full (step 232) . If the mass storage is not full, i.e., it has a block with its used/free flag unset, the algorithm continues and stores the data file into such a block (step 234) .
  • step 236 it is next determined whether there are any blocks which have both the old/new flag set AND the erase inhibit flag unset. If there are no blocks which have both the old/new flag set AND the erase inhibit flag unset (step 236) , the system of the present invention erases the data file, used/free flag and old/new flag in each block having its old/new flag set, and erases the counter and erase inhibit flag for every block (step 238) . Step 238 is also performed in the event there are insufficient blocks remaining to store a pending data file. The algorithm then returns to block (step 232) to determine whether the disk is full.
  • step 236 If the system can find a block having both the old/new flag set AND the erase inhibit flag unset (step 236) , then the system executes an erase procedure and erases the data file, used/free flag and old/new flag in each block having its old/new flag set. The counter is incremented and the erase inhibit flag for such blocks is not disturbed. It is then determined whether any block having its used/free flag unset has its counter at the maximum count (step 242) . If not, then the system of the present invention returns to decision step 232 and investigates again whether there is any block having its used/free flag unset (step 232) .
  • the erase inhibit flag is then set (step 244). Note that a data file will not be copied from a block having its erase count at one less than the maximum value, COUNT Max -l. Making such a reallocation from a source block having COUNT ⁇ -l to a destination block having COU T ⁇ results in having both blocks at COU T, ⁇ and no net gain. Further, the block previously having its erase count at COUNT ⁇ -1 is erased to no advantage, thus the erase cycle for that block would be wasted.
  • the old/new flag from the source block is then set (step 246) so that it can be erased during the next execution of an erase step 240. In that way the source block can be used for storage until its erase count reaches maximum and its erase inhibit flag is set.
  • the algorithm then returns to step 242 to determine whether there are now any blocks having an unset used/free flag with an erase count less than COUNT ⁇ . It will be understood that each time a data file is programmed or moved according to the algorithm of Figure 7 that the map in the CAM which correlates the logical address 308 to physical address 408 is updated so that the computer system can always access the data files.
  • a bit is programmed into the counter for each erase cycle rather than using binary counting.
  • an eight bit counter register would only be able to count to eight. This avoids having to erase the counter and then reprogramming it with an incremented value as would be necessary for binary counting. This is preferred because it avoids having to temporarily store the count value for all of the blocks being erased.
  • the counter registers need not be erased until all the blocks reach maximum count and there is a general erase. Because the mass storage apparatus of the present invention can accommodate large data storage, it is likely that many blocks will be flagged for a clean-out erase.
  • a read instruction is received by the mass storage apparatus of the present invention from the computer system (step 270) . Concurrent with receiving the read instruction, the system also receives the logical address 308 of the data file needed by the computer system (step 271) .
  • the apparatus of the present invention concatenates all the appropriate flags to the logical address 308 including having a set used/free flag, and unset new/old and defect flags (step 272) . If a match is found in the CAM (step 273) , the data file is read (step 275) otherwise a signal is returned to the computer system that the data file was not found (step 274) .

Abstract

A semiconductor mass storage system and architecture can be substituted for a rotating hard disk. The system and architecture avoid an erase cycle each time information stored in the mass storage is changed. (The erase cycle is understood to include, fully programming the block to be erased, and then erasing the block.) Erase cycles are avoided by programming an altered data file into an empty mass storage block rather than over itself as hard disk would. Periodically, the mass storage will need to be cleaned up. Secondly, means are provided for evenly using all blocks in the mass storage. These advantages are achieved through the use of several flags, a map (108) to correlate a logical address (308) of a block to a physical address (408) of that block and a count register for each block. In particular, flags (118, 112, 116) are provided for defective blocks, used blocks, old version of a block, and a count to determine the number of times a block has been erased and written and erase inhibit flag.

Description

FLASH MEMORY MASS STORAGE ARCHITECTURE
Field of the Invention
This invention relates to the field of mass storage for computers. More particularly, this invention relates to an architecture for replacing a hard disk with a semiconductor non-volatile memory and in particular flash memory.
Background of the Invention
Computers have used rotating magnetic media for mass storage of data, programs and information. Though widely used and commonly accepted, such hard disk drives suffer from a variety of deficiencies. Because of the rotation of the disk, there is an inherent latency in extracting information from a hard disk drive.
Other problems are especially dramatic in portable computers. In particular, hard disks are unable to withstand many of the kinds of physical shock that a portable computer will likely sustain. Further, the motor for rotating the disk consumes significant amounts of power decreasing the battery life for portable computers. Solid state memory is an ideal choice for replacing a hard disk drive for mass storage because it can resolve the problems cited above. Potential solutions have been proposed for replacing a hard disk drive with a semiconductor memory. For such a system to be truly useful, the memory must be non-volatile and alterable. The inventors have determined that flash memory is preferred for such a replacement. It should be noted that E2 PROM is also suitable as a replacement for a hard disk drive.
Flash memory is a single transistor memory cell which is programmable through hot electron injection and erasable through Fowler-Nordheim tunneling. The programming and erasing of such a memory cell requires current to pass through the dielectric surrounding a floating gate electrode. Because of this, such types of memory have a finite number of erase-write cycles. Eventually, the dielectric will fail. Manufacturers of flash cell devices specify the limit for the number erase-write cycles as between 10,000 and 100,000. Accordingly, unlike rotating magnetic media, a flash memory mass storage device does not have an indefinite lifetime.
Another requirement for a semiconductor mass storage device to be successful is that its use in lieu of a rotating media hard disk mass storage device be transparent to the system designer and the user. In other words, the designer of a computer incorporating such a semiconductor mass storage device could simply remove the hard disk and replace it with a semiconductor mass storage. All presently available commercial software should operate on a system employing such a semiconductor hard disk without the necessity of any modification.
SunDisk proposed an architecture for a semiconductor mass storage using flash memory at the Silicon Valley PC Design Conference July 9, 1991. That mass storage system included read-write block sizes of 512 Bytes (or multiples thereof) just like IBM PC compatible hard disk sector sizes. (IBM PC is a trademark of IBM Corporation.) During an erase cycle, an entire block is first fully programmed and then erased.
As in conventional hard disks, it appears in the SunDisk architecture that there is an erase-before-write cycle each time data is changed in the mass storage. Thus, if a program or data block is to be changed, the data is written to RAM and appropriately changed, the flash block is fully programmed, then erased and then reprogrammed to the new memory condition. Unlike a hard disk device, in a flash memory device an erase cycle is slow which can significantly reduce the performance of a system utilizing flash memory as its mass storage.
Though such an architecture provides a workable semiconductor mass storage, there are several inefficiencies. First of all, each time a memory block is changed, there is a delay to the entire system due to the necessary erase-before-write cycle before reprogramming the altered information back into the block. The overhead associated with erase-before-write cycles is costly in terras of system performance.
Secondly, hard disk users typically store both information which is rarely changed and information which is frequently changed. For example, a commercial spread sheet or word processing software programs stored on a user's system are rarely, if ever, changed. However, the spread sheet data files or word processing documents are frequently changed. Thus, different sectors of a hard disk typically have dramatically different usage in terms of the number of times the information stored thereon is changed. While this disparity has no impact on a hard disk because of its insensitivity to data changes, in a flash memory device, this variance can cause sections of the mass storage to wear out and be unusable significantly sooner than other sections of the mass storage.
Summary of the Invention
The present invention discloses two primary algorithms and an associated hardware architecture for a semiconductor mass storage device. It will be understood that data file in this patent document refers to any computer file including commercial software, a user program, word processing software document, spread sheet file and the like. The first algorithm provides means for avoiding an erase-before-write cycle when writing a modified data file back onto the mass storage device. Instead, no erase is performed and the modified data file is written onto an empty portion of the mass storage.. In addition, the second algorithm prevents any portion of the mass storage from being erased a substantially larger number of times than any other portion. This prevents any one block of the mass storage from failing and becoming unusable earlier than any other block thereby extending the life of the entire mass storage. The semiconductor mass storage architecture has blocks sized to conform with commercial hard disk sector sizes. The blocks are individually erasable. In one embodiment, the semiconductor mass storage of the present invention can be substituted for a rotating hard disk with no impact to the user, so that such a substitution will be transparent.
Means are provided for avoiding the erase-before-write cycle each time information stored in the mass storage is changed. (The erase cycle is understood to include, fully programming each bit in the block to be erased, and then erasing all the bits in the block.)
According to the first algorithm, erase cycles are avoided by programming an altered data file into an empty mass storage block rather than over itself after an erase cycle of that block as done on a conventional hard disk. This would ordinarily not be possible when using conventional mass storage because the central processor and commercial software available in conventional computer systems are not configured to track continually changing physical locations of data files. The present invention includes a programmable map to maintain a correlation between the logical address 308 and the physical address 408 of the updated information files.
Periodically, the mass storage will fill up because there have been no erase cycles. At such times, the mass storage needs to be cleaned up with a multi-sector erase as fully described in the detailed description below.
According to the second algorithm, means are provided for evenly using all blocks in the mass storage. A counter tracks the number of times each block is erased. A programmable maximum value for the counter is also provided. As the number of erase cycles for a block becomes one less than the maximum, the block is erased one last time and written with another file having a then smallest number of erase cycles. It is also prevented from being erased thereafter by setting its erase inhibit flag. After all blocks approach this maximum, all the erase counters and inhibit flags are cleared and the second algorithm is then repeated. In this way, no block can be erased a substantial number of times more than any other block.
These advantages are achieved through the use of several flags and a count register for each block. In particular, flags are provided for defective blocks, used blocks, old version of a block, a count to determine the number of times a block has been erased and written and an erase inhibit flag.
Brief Description of the Drawings
Figure 1 shows an architecture for a semiconductor mass storage. Figure 2 shows the architecture of Figure 1 wherein the data in one block has been altered and stored in a new physical address.
Figure 3 shows a block diagram of an erase cycle usage according to algorithm 1 of the present invention. Figure 4 shows a simplified block diagram of the old/new flag system integrally formed with the memory. Figure 5 shows a flow chart block diagram for algorithm 1.
Figure 6 shows an additional architecture according to the preferred embodiment of the present invention. Figure 7 shows a flow chart block diagram of algorithm 2 of the present invention.
Figure 8 shows a flow chart block diagram of a read algorithm according to the present invention.
Detailed Description of the Preferred Embodiment Figure 1 shows an architecture for a semiconductor mass storage according to the present invention. In the preferred embodiment, all of the memory storage is flash EEPROM. It is possible to substitute E2PROM for some or all of the data bits shown. A memory storage 100 is arranged into N blocks of data from zero through N-l. Each of the blocks of data is M Bytes long. In the preferred embodiment, each block is 512 Bytes long to correspond with a sector length in a commercially available hard disk drive. In addition to the memory data block 102, a flag 104 is directly related to each data block 102. The memory 100 can contain as much memory storage as a user desires. An example of a mass storage device might include 100 MByte of addressable storage.
A non-volatile content addressable memory (CAM) 106 is associated with the memory storage 100. In the preferred embodiment, the CAM 106 is formed of flash memory. The CAM 106 can also be E2PROM. There is one entry in the CAM 106 for every one of the N blocks in the mass storage 100. Each entry includes a number of fields which will be described below. The CAM 106 is also formed of a non-volatile memory because loss of its information would make retrieval of the data files stored in the mass storage 100 impossible. As described above in the Background of the Invention, conventional computer systems are not configured to track continually changing physical locations of data files. According to the present invention, each time a data file is changed it is stored into a new physical location in the mass storage. Thus, implementation of the architecture of the present invention requires a mapping of the logical address 308, i.e., the address where the computer system believes the data file is stored to the physical address 408, i.e., the actual location the data file can be found is stored in the mass storage.
The logical address 308 portion of the map 108 and the flags 112, 116 and 118 form part of the CAM 106. It is possible to use other storage means than a CAM to store the address map, such as a look-up table. However, a CAM is the most efficient means known to the inventors. It is not necessary that the physical address 408 portion of the map 108 form part of the CAM. Indeed, the physical address 408 portion of the- map 108 can be ordinary flash memory, E2PROM or even ROM. If ROM is selected for the physical address 408 array of the map 108, a defect in the ROM will prevent the block corresponding to that physical address 408 from ever being addressed. Accordingly, a changeable nonvolatile memory is preferred. Note that any replacement circuit for the CAM should be nonvolatile. Otherwise, loss or removal of power to the system will result in loss of the ability to find the data files in the mass storage.
Assume for example that a user is preparing a word processing document and instructs the computer to save the document. The document will be stored in the mass storage system as shown in Figure 1. The computer system will assign it a logical address 308, for example 526H. The mass storage system of the present invention will select a physical address 408 of an unused block or blocks in the mass storage 100 for storing the document, e.g. 728H. That map correlating the logical address 308 to the physical address 408 is stored in the CAM 106. As the data is programmed, the system of the present invention also sets the used/free flag 112 to indicate that this block has been written without being erased. The used/free flag 112 also forms a portion of the CAM 106. One used/free flag 112 is provided for each entry of the CAM 106.
Later, assume the user retrieves the document, makes a change and again instructs the computer to store the document. To avoid an erase-before-write cycle, the system of the present invention provides means for locating a block having its used/free flag 112 unset (not programmed) which indicates that the associated block is erased. The system then sets the used/free flag for the new block 114 (Figure 2) and then stores the modified document in that new block 114. Next, the system sets the old/new flag 116 of the previous version of the document indicating that this is an old unneeded version of the document. Lastly, the system updates the correlation between the logical address 308 and the actual physical address 408. In this way, the system of the present invention avoids the overhead of an erase cycle which is required in the erase-before-write of conventional systems to store a modified version of a previous document. The writing to mass storage process outlined above is repeated until the entire mass storage memory 100 has been filled. A full mass storage is indicated by no unset used/free flags 112 in the CAM 106. At that time a multi- sector erase is necessary and those blocks in the memory 100 and their associated CAM 106 entries having an old/new flag 116 set are all erased simultaneously. Note that it is not necessary for 100% of the blocks to have a set used/free flag 112 for a multi-sector erase to occur. For example, if a data file requiring three blocks were being written and only two blocks having unset used/free flags 112 were available a multi-sector erase can be run.
A simultaneous erase is not needed with prior art implementations because those embodiments utilize an erase- before-write cycle rather than retaining superseded versions of data files. In such circuits a latch of volatile logic circuits is set to couple the voltage necessary to erase the flash cells in the block. Because of the likely large number of memory blocks in the mass storage 100, if the CAM 106 and mass storage 100 are on the same integrated circuit (chip) coupling the old/new flag 116 to the latches in parallel would typically be very expensive in terms of surface area of the chip and coupling the old/new flags 116 serially to the latches would be expensive in terms of system performance. If the CAM 106 and the mass storage 100 are on separate chips, it is doubtful that either device could have sufficient I/O capability to interconnect the old/new flags 116 to the latches in parallel and thus, the system would suffer from a serial transfer of that information for a multi-sector erase.
Because of these problems it is preferable that no updating of the latches be performed prior to an erase of all blocks having a set old/new flag 116. To avoid this step, a plurality of old/new flag systems 104 are intimately associated with each block in the memory 102 and is programmed by the same sequence of instructions as the old/new flag 116 of the CAM 106.
Figure 4 shows a simplified block diagram of the old/new flag system 104 which includes a non-volatile bit 120 having data which mirrors the old/new flag 116. In addition there is a volatile latch 122 coupled to receive the data in the bit 120 from the latch during an erase cycle. At the time of an erase, the data in each of the bits 120 is simultaneously coupled to each appropriate ones of the latches 122 under control of a load signal coupled to each latch 122 over a load line L. Upon receiving a signal to perform the erase, the latch for every block having its associated bit 120 set then couples the voltage necessary to perform an erase of that block and its associated bit 120. After the erase is complete and verified, all the latches 122 are individually reset to a predetermined state under control of a reset signal coupled to each latch 122 over a reset line R. For certain applications of the present invention, especially for low power portable computers, a simultaneous erase of all blocks having their respective old/new flags set may be undesirable. For such applications, the blocks can be segregated into groups of blocks. Each group has a unique control line to load the latches from the nonvolatile bits. In this mode, during an erase cycle, the control lines are sequentially activated and the groups of blocks sequentially erased.
Figure 5 shows algorithm 1 according to the present invention. When the system of the present invention receives an instruction to program data into the mass storage (step 200) , then the system attempts to locate a free block (step 202), i.e., a block having an unset (not programmed) used/free flag. If successful, the system sets the used/free flag for that block and programs the data into that block (step 206) .
If on the other hand, the system is unable to locate a block having an unset used/free flag, the system erases the flags (used/free and old/new) and data for all blocks having a set old/new flag (step 204) and then searches for a block having an unset used/free flag (step 202) . Such a block has just been formed by step 204. The system then sets the used/free flag for that block and programs the data file into that block (step 206) . If the data file is a modified version of a previously existing file, the system must prevent the superseded version from being accessed. The.system determines whether the data file supersedes a previous data file (step 208) . If so, the system sets the old/new flag associated with the superseded block (step 210) . If on the other hand, the data file to be stored is a newly created data file, the step of setting the old/new flag (step 210) is skipped because there is no superseded block. Lastly, the map for correlating the logical address 308 to the physical address 408 is updated (step 212) .
By following the procedure outlined above, the overhead associated with an erase cycle is avoided for each write to the memory 100 except for periodically. This vastly improves the performance of the overall computer system employing the architecture of the present invention.
In the preferred embodiment of the present invention, the programming of the flash memory follows the procedure commonly understood by those of ordinary skill in the art. In other words, the program impulses are appropriately applied to the bits to be programmed and then compared to the data being programmed to ensure that proper programming has occurred. In the event that a bit fails to be erased or programmed properly, a defect flag 118 in the CAM 106 is set preventing that block from being used again.
In addition to saving the overhead of the erase cycle all but periodically, utilization of the present invention tends to more evenly distribute the erase cycles amongst certain portions of the blocks of the mass storage. Figure 3 schematically shows the types of information stored in utilizing a mass storage media 150. One portion of the mass storage 150 contains commercial applications software 152 such as word processing, spreadsheet, calendaring, calculators and the like. These portions of the mass storage 150 rarely, if ever, require an erase-reprogram cycle according to the algorithm described above.
A second section of the mass,storage 150 contains user data 154. The user data 154 is frequently altered requiring the information to be reprogrammed into blocks of the free space 156 under the algorithm described above. A third portion of the mass storage 150 contains free space 156 of unprogrammed blocks.
By following the algorithm above, the storage blocks in the portions 154 and 156 of the memory 150 will recycle data files and thus be erased and reprogrammed significantly more often than the commercial applications software portion 152 of the memory 150. Accordingly, the mass storage 150 will wear out more quickly in the user data 154 and the free space 156 sections of the memory requiring earlier replacement than in sections 152 of the mass storage having data files which are rarely changed. As the number of free blocks diminishes providing a smaller number of blocks through which to recycle data files, the remaining blocks become erased more frequently exacerbating the problem. A second algorithm is provided for leveling erase cycles amongst all the blocks within the entire mass storage device as shown in Figure 6. A counter is provided for each block to count the number of times each block has been erased and reprogrammed. An erase inhibit flag is also provided for each block. Once the erase count has reached the maximum for any block, the erase inhibit flag is set for that block. After that time that block cannot be erased until a clean-out erase is performed. Referring to Figure 3, if only algorithm 1 is used eventually all of the blocks in the user data 154 and the free space 156 portions of the mass storage 150 will reach the maximum count and have their respective erase inhibit flags set. Because of this, a reallocation of the rarely erased data files stored in the memory 152 is made into the memory 154 and/or 156. In this way, sections of the mass storage which have been erased numerous times are programmed with a reallocated data file which is rarely changed thereby allowing all sections of the mass storage to eventually approach parity of erase cycles. Like the multi-sector erase, a clean-out erase can be performed in the event that there is insufficient available storage for a data file presently being performed. For example, if all but two blocks have their respective erase inhibit flags set, and a three or more block data file is being programmed, a clean-out erase can be performed to provide sufficient storage for the data file.
Once the erase inhibit flag is set for all the blocks, indicating that all the blocks have achieved parity in erase cycles, the erase inhibit and erase count registers are erased and the cycle is repeated. The selection of the maximum count depends upon the system requirements. As the value for the maximum count increases, the disparity between erase count cycles of various blocks can also increase. However, because data is shifted as a result of achieving maximum erase count this process of smoothing cycles throughout the mass storage of itself introduces additional erase cycles because a block of information is transferred from a physical block having few erases to a block having the maximum number of erases. Accordingly, though low maximum count values reduce the disparity between erase cycles amongst the blocks it also increases the number of erase cycles to which the blocks are subjected. Accordingly, individual users may select an erase count depending upon the system needs.
In the preferred embodiment, algorithm 2 is merged with algorithm 1 as shown in Figure 7. An instruction is provided by the computer system to write a data file to the mass storage (step 230) which starts the combined algorithm 1 and algorithm 2 sequence. It is first determined whether the mass storage is full (step 232) . If the mass storage is not full, i.e., it has a block with its used/free flag unset, the algorithm continues and stores the data file into such a block (step 234) .
If on the other hand, it is determined that there are no free blocks, then it is next determined whether there are any blocks which have both the old/new flag set AND the erase inhibit flag unset (step 236) . If there are no blocks which have both the old/new flag set AND the erase inhibit flag unset (step 236) , the system of the present invention erases the data file, used/free flag and old/new flag in each block having its old/new flag set, and erases the counter and erase inhibit flag for every block (step 238) . Step 238 is also performed in the event there are insufficient blocks remaining to store a pending data file. The algorithm then returns to block (step 232) to determine whether the disk is full.
If the system can find a block having both the old/new flag set AND the erase inhibit flag unset (step 236) , then the system executes an erase procedure and erases the data file, used/free flag and old/new flag in each block having its old/new flag set. The counter is incremented and the erase inhibit flag for such blocks is not disturbed. It is then determined whether any block having its used/free flag unset has its counter at the maximum count (step 242) . If not, then the system of the present invention returns to decision step 232 and investigates again whether there is any block having its used/free flag unset (step 232) .
On the other hand, if there is a block having its erase count at the maximum value, a data file is copied from another block having the then least count value (step 244) into the location having COUNT = COUNTMax. The erase inhibit flag is then set (step 244). Note that a data file will not be copied from a block having its erase count at one less than the maximum value, COUNTMax-l. Making such a reallocation from a source block having COUNT^-l to a destination block having COU T^ results in having both blocks at COU T,^ and no net gain. Further, the block previously having its erase count at COUNT^-1 is erased to no advantage, thus the erase cycle for that block would be wasted. The old/new flag from the source block is then set (step 246) so that it can be erased during the next execution of an erase step 240. In that way the source block can be used for storage until its erase count reaches maximum and its erase inhibit flag is set. The algorithm then returns to step 242 to determine whether there are now any blocks having an unset used/free flag with an erase count less than COUNT^. It will be understood that each time a data file is programmed or moved according to the algorithm of Figure 7 that the map in the CAM which correlates the logical address 308 to physical address 408 is updated so that the computer system can always access the data files.
The efficiency of these algorithms has been tested by simulation. In the simulation it was assumed that the mass storage was 50% filled with data files that are not changed, 30% with data files that are routinely changed and 20% empty. Of the 30% of the data files that are routinely changed, % are rewritten 70% of the time, % are rewritten 25% of the time and % are rewritten 5% of the time. The simulation showed that the algorithm 1 improves the number of cycles until any block has reached failure by between six and seven times and algorithm 2 by approximately two times over the improvement gained using algorithm 1 alone. Depending upon the design criterion of a target system, it is possible to utilize either algorithm 1, algorithm 2 or the preferred merged algorithm. Algorithm 1 and the merged algorithm have been described above.
In the preferred embodiment, a bit is programmed into the counter for each erase cycle rather than using binary counting. Thus, an eight bit counter register would only be able to count to eight. This avoids having to erase the counter and then reprogramming it with an incremented value as would be necessary for binary counting. This is preferred because it avoids having to temporarily store the count value for all of the blocks being erased. By programming a bit for each, the counter registers need not be erased until all the blocks reach maximum count and there is a general erase. Because the mass storage apparatus of the present invention can accommodate large data storage, it is likely that many blocks will be flagged for a clean-out erase. Either a temporary volatile storage would be necessary for each block to store the previous count value prior to incrementing and reprogramming or the erase and updating of the counters would have to be done one after the other. One solution requires integrated circuit surface area and the other degrades performance. Note however, that if binary counting is desired the erase counter can be erased each time the block is erased and immediately reprogrammed. Because this will happen only during the periodic erase cycle described relative to the first algorithm some system designers may find this acceptable.
The read algorithm according to the present invention is shown in Figure 8. A read instruction is received by the mass storage apparatus of the present invention from the computer system (step 270) . Concurrent with receiving the read instruction, the system also receives the logical address 308 of the data file needed by the computer system (step 271) . The apparatus of the present invention concatenates all the appropriate flags to the logical address 308 including having a set used/free flag, and unset new/old and defect flags (step 272) . If a match is found in the CAM (step 273) , the data file is read (step 275) otherwise a signal is returned to the computer system that the data file was not found (step 274) .
The present invention is described relative to a preferred embodiment. Modifications or improvements which become apparent to one of ordinary skill in the art after reading this disclosure are deemed within the spirit and scope of this invention.

Claims

C L A I M SWhat is claimed is:
1. A non-volatile semiconductor mass storage device comprising: a. a plurality of non-volatile storage blocks, wherein each block is selectively programmable to store data and is selectively erasable and further wherein each block has a finite number of erase cycle lifetimes; and b. means for ensuring no block is subjected to more than a predetermined larger number of erase cycles than any other block.
2. The device according to claim 1 wherein the means for ensuring comprises a counter for each block for counting each erase cycle to which that block has been subjected.
3. The device according to claim 2 wherein the counter includes a plurality of bits programmed by binary counting.
4. The device according to claim 2 wherein the counter includes a plurality of bits programmed by sequentially programming each bit, one at a time, wherein each programmed bit represents a count of one.
5. The device according to claim 2 further comprising means for setting a maximum count value which applies to all blocks, and further an erase inhibit flag having a set condition and an unset condition for each block for preventing further erases to a block having its erase inhibit flag in the set condition.
6. The device according to claim 5 further comprising: - 19 - a. means for determining whether the counter for any erased block is equal to the maximum count value; b. means for finding a block having a count value less than the maximum count value; and c. means for moving the block having a count value less than the maximum count value into an erased block having its counter equal to the maximum count value.
7. The device according to claim 6 further comprising means for finding a block having a then least count value and moving that block into the erased block having its counter equal to the maximum count value.
8. The device according to claim 5 wherein the blocks, the counter and the flag are formed of flash memory cells.
9. The device according to claim 5 wherein the blocks, the counter and the flag are formed of E2PROM cells.
10. A non-volatile semiconductor mass storage device comprising: a. a plurality of non-volatile storage blocks, wherein each block is selectively programmable to store data and is selectively erasable and further wherein each block has a finite number of erase cycle lifetimes; and b. a controller to ensure no block is subjected to more than a predetermined larger number of erase cycles than any other block.
11. The device according to claim 10 wherein the controller comprises a counter for each block for counting each erase cycle to which that block has been subjected. - 20 -
12. The device according to claim 11 wherein the counter includes a plurality of bits programmed by binary counting.
13. The device according to claim 11 wherein the counter includes a plurality of bits programmed by sequentially programming each bit, one at a time, wherein each programmed bit represents a count of one.
14. The device according to claim 11 further comprising means for setting a maximum count value which applies to all blocks, and further an erase inhibit flag having a set condition and an unset condition for each block for preventing further erases to a block having its erase inhibit flag in the set condition.
15. The device according to claim 12 further comprising: a. a first circuit to determine whether the counter for any erased block is equal to the maximum count value; b. a searcher to find a block having a count value less than the maximum count value; and c. a second circuit to move the block having a count value less than the maximum count value into an erased block having its counter equal to the maximum count value.
16. The device according to claim 15 further comprising means for finding a block having a then least count value and moving that block into the erased block having its counter equal to the maximum count value.
17. The device according to claim 12 wherein the blocks, the counter and the flag are formed of flash memory cells. - 21 -
18. The device according to claim 12 wherein the blocks, the counter and the flag are formed of E2PROM cells.
19. A non-volatile semiconductor mass storage device comprising: a. a plurality of non-volatile storage blocks, wherein each block is selectively programmable to store data and is selectively erasable and further wherein each block has a finite number of erase cycle lifetimes; b. a plurality of counters, one counter for each block, to count each time its associated block is erased, wherein a programmable maximum value is applied to all the blocks; c. a plurality of erase inhibit flags, each flag having a set condition and an unset condition, one flag for each block, wherein the flag for a block having a count value equal to the maximum value is set to prevent further erasing of that block; d. a controller to switch data in a block having a relatively small count value into a block having the maximum value; and e. an erase element to clear all counters and all flags once all flags are set.
20. The device according to claim 19 further comprising: a. means for determining whether the counter for any erased block is equal to the maximum count value; b. means for finding a block having a count value less than the maximum count value; and c. means for moving the block having a count value less than the maximum count value into an erased block having its counter equal to the maximum count value.
21. The device according to claim 20 further comprising means for finding a block having a then least counter value and moving that block into the erased block having its counter equal to the maximum count value.
22. The device according to claim 21 wherein the blocks, the counter and the flag are formed of flash memory cells.
23. The device according to claim 21 wherein the blocks, the counter and the flag are formed of E2PROM cells.
24. A method of maintaining parity of erase cycles amongst storage blocks in a non-volatile semiconductor mass storage device having a plurality of non-volatile storage blocks, wherein each block is selectively programmable to store data and is selectively erasable and further wherein each block has a finite number of erase cycle lifetimes, the method comprising; a. counting and maintaining a count value for each block representing each time it is erased; b. setting a predetermined maximum value; c. preventing an erase cycle from being performed on a block having a count value equal to the maximum value; d. switching data from a block having a smallest count value into a block having the maximum value; and e. erasing the count values once all blocks have been erased the maximum value of times.
PCT/US1994/003154 1993-03-26 1994-03-23 Flashmemory mass storage architecture WO1994023432A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
EP94912849A EP0693216A4 (en) 1993-03-26 1994-03-23 Flashmemory mass storage architecture

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US08/038,668 1993-03-26
US08/038,668 US5388083A (en) 1993-03-26 1993-03-26 Flash memory mass storage architecture

Publications (1)

Publication Number Publication Date
WO1994023432A1 true WO1994023432A1 (en) 1994-10-13

Family

ID=21901226

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US1994/003154 WO1994023432A1 (en) 1993-03-26 1994-03-23 Flashmemory mass storage architecture

Country Status (4)

Country Link
US (1) US5388083A (en)
EP (1) EP0693216A4 (en)
CA (1) CA2161345A1 (en)
WO (1) WO1994023432A1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6069827A (en) * 1995-09-27 2000-05-30 Memory Corporation Plc Memory system
CN101409108B (en) * 2007-10-09 2011-04-13 群联电子股份有限公司 Average abrasion method and controller using the same
GB2490991A (en) * 2011-05-19 2012-11-21 Ibm Wear leveling in solid state memory

Families Citing this family (195)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0617363B1 (en) * 1989-04-13 2000-01-26 SanDisk Corporation Defective cell substitution in EEprom array
US6230233B1 (en) * 1991-09-13 2001-05-08 Sandisk Corporation Wear leveling techniques for flash EEPROM systems
EP0596198B1 (en) * 1992-07-10 2000-03-29 Sony Corporation Flash eprom with erase verification and address scrambling architecture
US5359570A (en) * 1992-11-13 1994-10-25 Silicon Storage Technology, Inc. Solid state peripheral storage device
US5581723A (en) * 1993-02-19 1996-12-03 Intel Corporation Method and apparatus for retaining flash block structure data during erase operations in a flash EEPROM memory array
US5592641A (en) * 1993-06-30 1997-01-07 Intel Corporation Method and device for selectively locking write access to blocks in a memory array using write protect inputs and block enabled status
US5559988A (en) * 1993-12-30 1996-09-24 Intel Corporation Method and circuitry for queuing snooping, prioritizing and suspending commands
KR950034271A (en) * 1994-01-26 1995-12-28 오오가 노리오 Nonvolatile Semiconductor Flash Memory
FR2718867B1 (en) * 1994-04-13 1996-05-24 Sgs Thomson Microelectronics Method for erasing a memory and implementation circuits.
US5473569A (en) * 1994-08-01 1995-12-05 Motorola, Inc. Method for operating a flash memory
JPH08124393A (en) * 1994-10-20 1996-05-17 Fujitsu Ltd Method for controlling memory having limitation on number of data erasing times and device using this method
JP2671860B2 (en) * 1995-03-30 1997-11-05 日本電気株式会社 File system for flash memory
US5568423A (en) * 1995-04-14 1996-10-22 Unisys Corporation Flash memory wear leveling system providing immediate direct access to microprocessor
US5907856A (en) * 1995-07-31 1999-05-25 Lexar Media, Inc. Moving sectors within a block of information in a flash memory mass storage architecture
US5845313A (en) 1995-07-31 1998-12-01 Lexar Direct logical block addressing flash memory mass storage architecture
US6978342B1 (en) * 1995-07-31 2005-12-20 Lexar Media, Inc. Moving sectors within a block of information in a flash memory mass storage architecture
US5930815A (en) * 1995-07-31 1999-07-27 Lexar Media, Inc. Moving sequential sectors within a block of information in a flash memory mass storage architecture
US6801979B1 (en) 1995-07-31 2004-10-05 Lexar Media, Inc. Method and apparatus for memory control circuit
US6757800B1 (en) 1995-07-31 2004-06-29 Lexar Media, Inc. Increasing the memory performance of flash memory devices by writing sectors simultaneously to multiple flash memory devices
US8171203B2 (en) * 1995-07-31 2012-05-01 Micron Technology, Inc. Faster write operations to nonvolatile memory using FSInfo sector manipulation
US6081878A (en) * 1997-03-31 2000-06-27 Lexar Media, Inc. Increasing the memory performance of flash memory devices by writing sectors simultaneously to multiple flash memory devices
US5838614A (en) 1995-07-31 1998-11-17 Lexar Microsystems, Inc. Identification and verification of a sector within a block of mass storage flash memory
US6728851B1 (en) 1995-07-31 2004-04-27 Lexar Media, Inc. Increasing the memory performance of flash memory devices by writing sectors simultaneously to multiple flash memory devices
US5835935A (en) * 1995-09-13 1998-11-10 Lexar Media, Inc. Method of and architecture for controlling system data with automatic wear leveling in a semiconductor non-volatile mass storage memory
US6125435A (en) * 1995-09-13 2000-09-26 Lexar Media, Inc. Alignment of cluster address to block addresses within a semiconductor non-volatile mass storage memory
JP2848300B2 (en) * 1995-12-27 1999-01-20 日本電気株式会社 Nonvolatile semiconductor memory device
EP0806773B1 (en) * 1996-05-09 2003-03-19 STMicroelectronics S.r.l. Electrically erasable and programmable non-volatile memory device with testable redundancy circuits
US5745418A (en) * 1996-11-25 1998-04-28 Macronix International Co., Ltd. Flash memory mass storage system
US5956473A (en) * 1996-11-25 1999-09-21 Macronix International Co., Ltd. Method and system for managing a flash memory mass storage system
US5928370A (en) * 1997-02-05 1999-07-27 Lexar Media, Inc. Method and apparatus for verifying erasure of memory blocks within a non-volatile memory structure
US6411546B1 (en) 1997-03-31 2002-06-25 Lexar Media, Inc. Nonvolatile memory using flexible erasing methods and method and system for using same
US6034897A (en) * 1999-04-01 2000-03-07 Lexar Media, Inc. Space management for managing high capacity nonvolatile memory
US6122195A (en) * 1997-03-31 2000-09-19 Lexar Media, Inc. Method and apparatus for decreasing block write operation times performed on nonvolatile memory
US5956743A (en) * 1997-08-25 1999-09-21 Bit Microsystems, Inc. Transparent management at host interface of flash-memory overhead-bytes using flash-specific DMA having programmable processor-interrupt of high-level operations
US6000006A (en) * 1997-08-25 1999-12-07 Bit Microsystems, Inc. Unified re-map and cache-index table with dual write-counters for wear-leveling of non-volatile flash RAM mass storage
US5822251A (en) * 1997-08-25 1998-10-13 Bit Microsystems, Inc. Expandable flash-memory mass-storage using shared buddy lines and intermediate flash-bus between device-specific buffers and flash-intelligent DMA controllers
US5848026A (en) * 1997-12-08 1998-12-08 Atmel Corporation Integrated circuit with flag register for block selection of nonvolatile cells for bulk operations
US6076137A (en) 1997-12-11 2000-06-13 Lexar Media, Inc. Method and apparatus for storing location identification information within non-volatile memory devices
US6400634B1 (en) 1997-12-23 2002-06-04 Macronix International Co., Ltd. Technique for increasing endurance of integrated circuit memory
WO1999033057A1 (en) * 1997-12-23 1999-07-01 Macronix International Co., Ltd. Technique for increasing endurance of integrated circuit memory
GB9801373D0 (en) 1998-01-22 1998-03-18 Memory Corp Plc Memory system
JP3714969B2 (en) * 1998-03-02 2005-11-09 レクサー・メディア・インコーポレイテッド Flash memory card with improved operating mode detection and user-friendly interfacing system
US6040997A (en) * 1998-03-25 2000-03-21 Lexar Media, Inc. Flash memory leveling architecture having no external latch
KR100342977B1 (en) * 1998-09-28 2002-09-18 주식회사 하이닉스반도체 Control circuit of address in a semiconductor memory device
US6901457B1 (en) 1998-11-04 2005-05-31 Sandisk Corporation Multiple mode communications system
AU1729100A (en) 1998-11-17 2000-06-05 Lexar Media, Inc. Method and apparatus for memory control circuit
WO2000060605A1 (en) 1999-04-01 2000-10-12 Lexar Media, Inc. Space management for managing high capacity nonvolatile memory
US6141249A (en) * 1999-04-01 2000-10-31 Lexar Media, Inc. Organization of blocks within a nonvolatile memory unit to effectively decrease sector write operation time
JP2000285688A (en) * 1999-04-01 2000-10-13 Mitsubishi Electric Corp Non-volatile semiconductor memory
US6148354A (en) * 1999-04-05 2000-11-14 M-Systems Flash Disk Pioneers Ltd. Architecture for a universal serial bus-based PC flash disk
CN1088218C (en) * 1999-11-14 2002-07-24 邓国顺 Electronic flash storage method and device for data processing system
JP4842417B2 (en) * 1999-12-16 2011-12-21 ソニー株式会社 Recording device
JP4165990B2 (en) * 1999-12-20 2008-10-15 Tdk株式会社 MEMORY CONTROLLER, FLASH MEMORY SYSTEM PROVIDED WITH MEMORY CONTROLLER, AND METHOD FOR WRITEING DATA TO FLASH MEMORY
US7102671B1 (en) 2000-02-08 2006-09-05 Lexar Media, Inc. Enhanced compact flash memory card
US6426893B1 (en) 2000-02-17 2002-07-30 Sandisk Corporation Flash eeprom system with simultaneous multiple data sector programming and storage of physical block characteristics in other designated blocks
US6721843B1 (en) 2000-07-07 2004-04-13 Lexar Media, Inc. Flash memory architecture implementing simultaneously programmable multiple flash memory banks that are host compatible
US6567307B1 (en) 2000-07-21 2003-05-20 Lexar Media, Inc. Block management for mass storage
US7167944B1 (en) 2000-07-21 2007-01-23 Lexar Media, Inc. Block management for mass storage
US7155559B1 (en) 2000-08-25 2006-12-26 Lexar Media, Inc. Flash memory architecture with separate storage of overhead and user data
US6772274B1 (en) 2000-09-13 2004-08-03 Lexar Media, Inc. Flash memory system and method implementing LBA to PBA correlation within flash memory array
US6473845B1 (en) * 2000-09-28 2002-10-29 Hewlett-Packard Company System and method for dynamically updating memory address mappings
US6763424B2 (en) 2001-01-19 2004-07-13 Sandisk Corporation Partial block data programming and reading operations in a non-volatile memory
US6549467B2 (en) * 2001-03-09 2003-04-15 Micron Technology, Inc. Non-volatile memory device with erase address register
US6732221B2 (en) 2001-06-01 2004-05-04 M-Systems Flash Disk Pioneers Ltd Wear leveling of static areas in flash memory
KR100389867B1 (en) 2001-06-04 2003-07-04 삼성전자주식회사 Flash memory management method
US7108975B2 (en) * 2001-09-21 2006-09-19 Regents Of The University Of Michigan Atlastin
GB0123421D0 (en) * 2001-09-28 2001-11-21 Memquest Ltd Power management system
GB0123416D0 (en) * 2001-09-28 2001-11-21 Memquest Ltd Non-volatile memory control
GB0123412D0 (en) 2001-09-28 2001-11-21 Memquest Ltd Memory system sectors
GB0123415D0 (en) * 2001-09-28 2001-11-21 Memquest Ltd Method of writing data to non-volatile memory
GB0123417D0 (en) * 2001-09-28 2001-11-21 Memquest Ltd Improved data processing
GB0123410D0 (en) 2001-09-28 2001-11-21 Memquest Ltd Memory system for data storage and retrieval
GB0123419D0 (en) 2001-09-28 2001-11-21 Memquest Ltd Data handling system
DE10163342A1 (en) * 2001-12-21 2003-07-10 Elektro Beckhoff Gmbh Unterneh Data transmission method, serial bus system and connection unit for a passive bus device
US6950918B1 (en) 2002-01-18 2005-09-27 Lexar Media, Inc. File management of one-time-programmable nonvolatile memory devices
US6957295B1 (en) 2002-01-18 2005-10-18 Lexar Media, Inc. File management of one-time-programmable nonvolatile memory devices
US7231643B1 (en) 2002-02-22 2007-06-12 Lexar Media, Inc. Image rescue system including direct communication between an application program and a device driver
US7277011B2 (en) * 2002-02-22 2007-10-02 Micron Technology, Inc. Removable memory media with integral indicator light
JP3833970B2 (en) * 2002-06-07 2006-10-18 株式会社東芝 Nonvolatile semiconductor memory
US6973531B1 (en) 2002-10-28 2005-12-06 Sandisk Corporation Tracking the most frequently erased blocks in non-volatile memory systems
US7181611B2 (en) * 2002-10-28 2007-02-20 Sandisk Corporation Power management block for use in a non-volatile memory system
US6831865B2 (en) * 2002-10-28 2004-12-14 Sandisk Corporation Maintaining erase counts in non-volatile storage systems
US8412879B2 (en) * 2002-10-28 2013-04-02 Sandisk Technologies Inc. Hybrid implementation for error correction codes within a non-volatile memory system
US7234036B1 (en) 2002-10-28 2007-06-19 Sandisk Corporation Method and apparatus for resolving physical blocks associated with a common logical block
US7171536B2 (en) * 2002-10-28 2007-01-30 Sandisk Corporation Unusable block management within a non-volatile memory system
US7526599B2 (en) * 2002-10-28 2009-04-28 Sandisk Corporation Method and apparatus for effectively enabling an out of sequence write process within a non-volatile memory system
US7035967B2 (en) * 2002-10-28 2006-04-25 Sandisk Corporation Maintaining an average erase count in a non-volatile storage system
US7096313B1 (en) 2002-10-28 2006-08-22 Sandisk Corporation Tracking the least frequently erased blocks in non-volatile memory systems
US20040083334A1 (en) * 2002-10-28 2004-04-29 Sandisk Corporation Method and apparatus for managing the integrity of data in non-volatile memory system
US7254668B1 (en) 2002-10-28 2007-08-07 Sandisk Corporation Method and apparatus for grouping pages within a block
US7174440B2 (en) * 2002-10-28 2007-02-06 Sandisk Corporation Method and apparatus for performing block caching in a non-volatile memory system
CN100483552C (en) * 2002-10-28 2009-04-29 桑迪士克股份有限公司 Automated wear leveling in non-volatile storage systems
US7039788B1 (en) 2002-10-28 2006-05-02 Sandisk Corporation Method and apparatus for splitting a logical block
US6985992B1 (en) 2002-10-28 2006-01-10 Sandisk Corporation Wear-leveling in non-volatile storage systems
US7103732B1 (en) * 2002-10-28 2006-09-05 Sandisk Corporation Method and apparatus for managing an erase count block
US7296043B2 (en) * 2003-05-30 2007-11-13 Microsoft Corporation Memory file size adjustment
US6973519B1 (en) 2003-06-03 2005-12-06 Lexar Media, Inc. Card identification compatibility
US7559004B1 (en) 2003-10-01 2009-07-07 Sandisk Corporation Dynamic redundant area configuration in a non-volatile memory system
US7188228B1 (en) * 2003-10-01 2007-03-06 Sandisk Corporation Hybrid mapping implementation within a non-volatile memory system
US7089349B2 (en) * 2003-10-28 2006-08-08 Sandisk Corporation Internal maintenance schedule request for non-volatile memory system
US7032087B1 (en) 2003-10-28 2006-04-18 Sandisk Corporation Erase count differential table within a non-volatile memory system
US8706990B2 (en) 2003-10-28 2014-04-22 Sandisk Technologies Inc. Adaptive internal table backup for non-volatile memory system
JP2007515024A (en) 2003-12-17 2007-06-07 レクサー メディア, インコーポレイテッド Activation of electronic device sales locations to avoid theft
WO2005081891A2 (en) * 2004-02-23 2005-09-09 Lexar Media, Inc. Secure compact flash
US7725628B1 (en) 2004-04-20 2010-05-25 Lexar Media, Inc. Direct secondary device interface by a host
US7370166B1 (en) 2004-04-30 2008-05-06 Lexar Media, Inc. Secure portable storage device
US7464306B1 (en) 2004-08-27 2008-12-09 Lexar Media, Inc. Status of overall health of nonvolatile memory
US7594063B1 (en) * 2004-08-27 2009-09-22 Lexar Media, Inc. Storage capacity status
US7441067B2 (en) * 2004-11-15 2008-10-21 Sandisk Corporation Cyclic flash memory wear leveling
US7315917B2 (en) * 2005-01-20 2008-01-01 Sandisk Corporation Scheduling of housekeeping operations in flash memory systems
US20060161724A1 (en) * 2005-01-20 2006-07-20 Bennett Alan D Scheduling of housekeeping operations in flash memory systems
US9104315B2 (en) 2005-02-04 2015-08-11 Sandisk Technologies Inc. Systems and methods for a mass data storage system having a file-based interface to a host and a non-file-based interface to secondary storage
US20060184719A1 (en) * 2005-02-16 2006-08-17 Sinclair Alan W Direct data file storage implementation techniques in flash memories
US20060184718A1 (en) * 2005-02-16 2006-08-17 Sinclair Alan W Direct file data programming and deletion in flash memories
US7877539B2 (en) * 2005-02-16 2011-01-25 Sandisk Corporation Direct data file storage in flash memories
US7224604B2 (en) * 2005-03-14 2007-05-29 Sandisk Il Ltd. Method of achieving wear leveling in flash memory using relative grades
US7984084B2 (en) * 2005-08-03 2011-07-19 SanDisk Technologies, Inc. Non-volatile memory with scheduled reclaim operations
US7552271B2 (en) 2005-08-03 2009-06-23 Sandisk Corporation Nonvolatile memory with block management
US7949845B2 (en) * 2005-08-03 2011-05-24 Sandisk Corporation Indexing of file data in reprogrammable non-volatile memories that directly store data files
US7558906B2 (en) 2005-08-03 2009-07-07 Sandisk Corporation Methods of managing blocks in nonvolatile memory
US7669003B2 (en) * 2005-08-03 2010-02-23 Sandisk Corporation Reprogrammable non-volatile memory systems with indexing of directly stored data files
US7480766B2 (en) * 2005-08-03 2009-01-20 Sandisk Corporation Interfacing systems operating through a logical address space and on a direct data file basis
US7627733B2 (en) * 2005-08-03 2009-12-01 Sandisk Corporation Method and system for dual mode access for storage devices
US7752382B2 (en) * 2005-09-09 2010-07-06 Sandisk Il Ltd Flash memory storage system and method
JP4751163B2 (en) * 2005-09-29 2011-08-17 株式会社東芝 Memory system
US7529905B2 (en) 2005-10-13 2009-05-05 Sandisk Corporation Method of storing transformed units of data in a memory system having fixed sized storage blocks
US7814262B2 (en) * 2005-10-13 2010-10-12 Sandisk Corporation Memory system storing transformed units of data in fixed sized storage blocks
US7877540B2 (en) * 2005-12-13 2011-01-25 Sandisk Corporation Logically-addressed file storage methods
US7644251B2 (en) * 2005-12-19 2010-01-05 Sigmatel, Inc. Non-volatile solid-state memory controller
US7769978B2 (en) 2005-12-21 2010-08-03 Sandisk Corporation Method and system for accessing non-volatile storage devices
US7793068B2 (en) * 2005-12-21 2010-09-07 Sandisk Corporation Dual mode access for non-volatile storage devices
US7747837B2 (en) * 2005-12-21 2010-06-29 Sandisk Corporation Method and system for accessing non-volatile storage devices
US7711890B2 (en) 2006-06-06 2010-05-04 Sandisk Il Ltd Cache control in a non-volatile memory device
US20080091901A1 (en) * 2006-10-12 2008-04-17 Alan David Bennett Method for non-volatile memory with worst-case control data management
US20080091871A1 (en) * 2006-10-12 2008-04-17 Alan David Bennett Non-volatile memory with worst-case control data management
TWI326028B (en) * 2006-11-20 2010-06-11 Silicon Motion Inc Method for flash memory data management
US9116823B2 (en) 2006-12-06 2015-08-25 Intelligent Intellectual Property Holdings 2 Llc Systems and methods for adaptive error-correction coding
US8074011B2 (en) * 2006-12-06 2011-12-06 Fusion-Io, Inc. Apparatus, system, and method for storage space recovery after reaching a read count limit
US8495292B2 (en) * 2006-12-06 2013-07-23 Fusion-Io, Inc. Apparatus, system, and method for an in-server storage area network
US9495241B2 (en) 2006-12-06 2016-11-15 Longitude Enterprise Flash S.A.R.L. Systems and methods for adaptive data storage
US8140737B2 (en) * 2007-01-30 2012-03-20 Skymedi Corporation Method for enhancing life cycle of memory
JP2008257773A (en) * 2007-04-02 2008-10-23 Toshiba Corp Nonvolatile semiconductor memory device, method for controlling the same, nonvolatile semiconductor memory system, and memory card
US8959307B1 (en) 2007-11-16 2015-02-17 Bitmicro Networks, Inc. Reduced latency memory read transactions in storage devices
US8195912B2 (en) * 2007-12-06 2012-06-05 Fusion-io, Inc Apparatus, system, and method for efficient mapping of virtual and physical addresses
US7836226B2 (en) 2007-12-06 2010-11-16 Fusion-Io, Inc. Apparatus, system, and method for coordinating storage requests in a multi-processor/multi-thread environment
WO2009156877A1 (en) * 2008-06-24 2009-12-30 Sandisk Il Ltd. Method and apparatus for error correction according to erase counts of a solid-state memory
US8041886B2 (en) * 2008-09-15 2011-10-18 Seagate Technology Llc System and method of managing memory
US8094500B2 (en) * 2009-01-05 2012-01-10 Sandisk Technologies Inc. Non-volatile memory and method with write cache partitioning
US8244960B2 (en) * 2009-01-05 2012-08-14 Sandisk Technologies Inc. Non-volatile memory and method with write cache partition management methods
US8040744B2 (en) * 2009-01-05 2011-10-18 Sandisk Technologies Inc. Spare block management of non-volatile memories
US20100174845A1 (en) * 2009-01-05 2010-07-08 Sergey Anatolievich Gorobets Wear Leveling for Non-Volatile Memories: Maintenance of Experience Count and Passive Techniques
US8700840B2 (en) * 2009-01-05 2014-04-15 SanDisk Technologies, Inc. Nonvolatile memory with write cache having flush/eviction methods
US8276042B2 (en) 2009-02-03 2012-09-25 Micron Technology, Inc. Determining sector status in a memory device
US8665601B1 (en) 2009-09-04 2014-03-04 Bitmicro Networks, Inc. Solid state drive with improved enclosure assembly
US9135190B1 (en) 2009-09-04 2015-09-15 Bitmicro Networks, Inc. Multi-profile memory controller for computing devices
US8447908B2 (en) 2009-09-07 2013-05-21 Bitmicro Networks, Inc. Multilevel memory bus system for solid-state mass storage
US8560804B2 (en) 2009-09-14 2013-10-15 Bitmicro Networks, Inc. Reducing erase cycles in an electronic storage device that uses at least one erase-limited memory device
US9110594B2 (en) * 2009-11-04 2015-08-18 Seagate Technology Llc File management system for devices containing solid-state media
US9396104B1 (en) 2010-03-22 2016-07-19 Seagate Technology, Llc Accessing compressed data of varying-sized quanta in non-volatile memory
US8949506B2 (en) 2010-07-30 2015-02-03 Apple Inc. Initiating wear leveling for a non-volatile memory
US8521948B2 (en) 2011-01-03 2013-08-27 Apple Inc. Handling dynamic and static data for a system having non-volatile memory
US9372755B1 (en) 2011-10-05 2016-06-21 Bitmicro Networks, Inc. Adaptive power cycle sequences for data recovery
US9043669B1 (en) 2012-05-18 2015-05-26 Bitmicro Networks, Inc. Distributed ECC engine for storage media
US8910017B2 (en) 2012-07-02 2014-12-09 Sandisk Technologies Inc. Flash memory with random partition
US8750045B2 (en) 2012-07-27 2014-06-10 Sandisk Technologies Inc. Experience count dependent program algorithm for flash memory
US9423457B2 (en) 2013-03-14 2016-08-23 Bitmicro Networks, Inc. Self-test solution for delay locked loops
US10489318B1 (en) 2013-03-15 2019-11-26 Bitmicro Networks, Inc. Scatter-gather approach for parallel data transfer in a mass storage system
US9916213B1 (en) 2013-03-15 2018-03-13 Bitmicro Networks, Inc. Bus arbitration with routing and failover mechanism
US9672178B1 (en) 2013-03-15 2017-06-06 Bitmicro Networks, Inc. Bit-mapped DMA transfer with dependency table configured to monitor status so that a processor is not rendered as a bottleneck in a system
US9875205B1 (en) 2013-03-15 2018-01-23 Bitmicro Networks, Inc. Network of memory systems
US9734067B1 (en) 2013-03-15 2017-08-15 Bitmicro Networks, Inc. Write buffering
US9400617B2 (en) 2013-03-15 2016-07-26 Bitmicro Networks, Inc. Hardware-assisted DMA transfer with dependency table configured to permit-in parallel-data drain from cache without processor intervention when filled or drained
US9934045B1 (en) 2013-03-15 2018-04-03 Bitmicro Networks, Inc. Embedded system boot from a storage device
US9720603B1 (en) 2013-03-15 2017-08-01 Bitmicro Networks, Inc. IOC to IOC distributed caching architecture
US9842024B1 (en) 2013-03-15 2017-12-12 Bitmicro Networks, Inc. Flash electronic disk with RAID controller
US9971524B1 (en) 2013-03-15 2018-05-15 Bitmicro Networks, Inc. Scatter-gather approach for parallel data transfer in a mass storage system
US9501436B1 (en) 2013-03-15 2016-11-22 Bitmicro Networks, Inc. Multi-level message passing descriptor
US9430386B2 (en) 2013-03-15 2016-08-30 Bitmicro Networks, Inc. Multi-leveled cache management in a hybrid storage system
US9798688B1 (en) 2013-03-15 2017-10-24 Bitmicro Networks, Inc. Bus arbitration with routing and failover mechanism
US10120694B2 (en) 2013-03-15 2018-11-06 Bitmicro Networks, Inc. Embedded system boot from a storage device
US20150143197A1 (en) * 2013-07-10 2015-05-21 Shmuel T. Klein Codes for Enhancing the Repeated Use of Flash Memory
US9952991B1 (en) 2014-04-17 2018-04-24 Bitmicro Networks, Inc. Systematic method on queuing of descriptors for multiple flash intelligent DMA engine operation
US10042792B1 (en) 2014-04-17 2018-08-07 Bitmicro Networks, Inc. Method for transferring and receiving frames across PCI express bus for SSD device
US10055150B1 (en) 2014-04-17 2018-08-21 Bitmicro Networks, Inc. Writing volatile scattered memory metadata to flash device
US10078604B1 (en) 2014-04-17 2018-09-18 Bitmicro Networks, Inc. Interrupt coalescing
US10025736B1 (en) 2014-04-17 2018-07-17 Bitmicro Networks, Inc. Exchange message protocol message transmission between two devices
US9811461B1 (en) 2014-04-17 2017-11-07 Bitmicro Networks, Inc. Data storage system
US10338817B2 (en) 2014-12-30 2019-07-02 Sandisk Technologies Llc Systems and methods for storage recovery
US10552050B1 (en) 2017-04-07 2020-02-04 Bitmicro Llc Multi-dimensional computer storage system
US10740181B2 (en) 2018-03-06 2020-08-11 Western Digital Technologies, Inc. Failed storage device rebuild method
US10860446B2 (en) * 2018-04-26 2020-12-08 Western Digital Technologiies, Inc. Failed storage device rebuild using dynamically selected locations in overprovisioned space
US10901848B2 (en) 2018-08-03 2021-01-26 Western Digital Technologies, Inc. Storage systems with peer data recovery
US10831603B2 (en) 2018-08-03 2020-11-10 Western Digital Technologies, Inc. Rebuild assist using failed storage device
US10824526B2 (en) 2018-08-03 2020-11-03 Western Digital Technologies, Inc. Using failed storage device in peer-to-peer storage system to perform storage-centric task
US11182258B2 (en) 2019-01-04 2021-11-23 Western Digital Technologies, Inc. Data rebuild using dynamic peer work allocation

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5034926A (en) * 1988-08-12 1991-07-23 Kabushiki Kaisha Toshiba Non-volatile semiconductor memory
US5138580A (en) * 1989-06-06 1992-08-11 Gemplus Card International Method for the erasure of memory cells, device designed to implement it, and use of said method in a device with non-supplied memory
JPH04332999A (en) * 1991-05-07 1992-11-19 Hitachi Koki Co Ltd Method of using memory
US5270979A (en) * 1991-03-15 1993-12-14 Sundisk Corporation Method for optimum erasing of EEPROM

Family Cites Families (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPS5945695A (en) * 1982-09-07 1984-03-14 Fujitsu Ltd Ic memory
JP2685173B2 (en) * 1986-05-31 1997-12-03 キヤノン株式会社 Memory write control method
JPH07109717B2 (en) * 1986-05-31 1995-11-22 キヤノン株式会社 Memory write control method
JPH0778997B2 (en) * 1987-10-30 1995-08-23 株式会社東芝 Non-volatile semiconductor memory
US5053990A (en) * 1988-02-17 1991-10-01 Intel Corporation Program/erase selection for flash memory
US5043940A (en) * 1988-06-08 1991-08-27 Eliyahou Harari Flash EEPROM memory systems having multistate storage cells
US5095344A (en) * 1988-06-08 1992-03-10 Eliyahou Harari Highly compact eprom and flash eeprom devices
US5168465A (en) * 1988-06-08 1992-12-01 Eliyahou Harari Highly compact EPROM and flash EEPROM devices
JPH0283892A (en) * 1988-09-20 1990-03-23 Fujitsu Ltd Semiconductor memory device
US5341368A (en) * 1992-05-05 1994-08-23 American Tel-A-Systems, Inc. Digital switching system interconnecting buses with imcompatible protocols
US5172338B1 (en) * 1989-04-13 1997-07-08 Sandisk Corp Multi-state eeprom read and write circuits and techniques
US5163021A (en) * 1989-04-13 1992-11-10 Sundisk Corporation Multi-state EEprom read and write circuits and techniques
US5065364A (en) * 1989-09-15 1991-11-12 Intel Corporation Apparatus for providing block erasing in a flash EPROM
US5200959A (en) * 1989-10-17 1993-04-06 Sundisk Corporation Device and method for defect handling in semi-conductor memory
US5134589A (en) * 1989-10-30 1992-07-28 Mitsubishi Denki Kabushiki Kaisha Semiconductor memory device having a flash write function
US5303198A (en) * 1990-09-28 1994-04-12 Fuji Photo Film Co., Ltd. Method of recording data in memory card having EEPROM and memory card system using the same
DE69021732T2 (en) * 1990-12-04 1996-01-18 Hewlett Packard Ltd Reprogrammable data storage system.
US5222109A (en) * 1990-12-28 1993-06-22 Ibm Corporation Endurance management for solid state files
US5272669A (en) * 1991-02-20 1993-12-21 Sundisk Corporation Method and structure for programming floating gate memory cells
DE4112292A1 (en) * 1991-04-15 1992-10-22 Nec Deutschland FIXED MASS STORAGE FOR A DATA PROCESSING SYSTEM, HOUSING FOR RECORDING THE FIXED MASS STORAGE AND DATA PROCESSING SYSTEM
EP0509184A1 (en) * 1991-04-18 1992-10-21 STMicroelectronics S.r.l. Flash memory erasable by sectors and related writing process
IT1254937B (en) * 1991-05-06 1995-10-11 DYNAMIC UPDATE OF NON-VOLATILE MEMORY IN A COMPUTER SYSTEM
DE4215063C2 (en) * 1991-05-10 1999-11-25 Intel Corp Device and method for changing pages in a non-volatile memory
JP2582487B2 (en) * 1991-07-12 1997-02-19 インターナショナル・ビジネス・マシーンズ・コーポレイション External storage system using semiconductor memory and control method thereof
US5341339A (en) * 1992-10-30 1994-08-23 Intel Corporation Method for wear leveling in a flash EEPROM memory

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5034926A (en) * 1988-08-12 1991-07-23 Kabushiki Kaisha Toshiba Non-volatile semiconductor memory
US5138580A (en) * 1989-06-06 1992-08-11 Gemplus Card International Method for the erasure of memory cells, device designed to implement it, and use of said method in a device with non-supplied memory
US5270979A (en) * 1991-03-15 1993-12-14 Sundisk Corporation Method for optimum erasing of EEPROM
JPH04332999A (en) * 1991-05-07 1992-11-19 Hitachi Koki Co Ltd Method of using memory

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP0693216A4 *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6069827A (en) * 1995-09-27 2000-05-30 Memory Corporation Plc Memory system
CN101409108B (en) * 2007-10-09 2011-04-13 群联电子股份有限公司 Average abrasion method and controller using the same
GB2490991A (en) * 2011-05-19 2012-11-21 Ibm Wear leveling in solid state memory
US9218277B2 (en) 2011-05-19 2015-12-22 International Business Machines Corporation Wear leveling
US9274944B2 (en) 2011-05-19 2016-03-01 International Business Machines Corporation Wear leveling
GB2490991B (en) * 2011-05-19 2017-08-30 Ibm Wear leveling

Also Published As

Publication number Publication date
US5388083A (en) 1995-02-07
EP0693216A1 (en) 1996-01-24
CA2161345A1 (en) 1994-10-13
EP0693216A4 (en) 1997-02-26

Similar Documents

Publication Publication Date Title
US5388083A (en) Flash memory mass storage architecture
US5479638A (en) Flash memory mass storage architecture incorporation wear leveling technique
US5485595A (en) Flash memory mass storage architecture incorporating wear leveling technique without using cam cells
US6912618B2 (en) Direct logical block addressing flash memory mass storage architecture
US8793430B2 (en) Electronic system having memory with a physical block having a sector storing data and indicating a move status of another sector of the physical block
US6223308B1 (en) Identification and verification of a sector within a block of mass STO rage flash memory

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): CA JP KR VN

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): AT BE CH DE DK ES FR GB GR IE IT LU MC NL PT SE

DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
121 Ep: the epo has been informed by wipo that ep was designated in this application
WWE Wipo information: entry into national phase

Ref document number: 2161345

Country of ref document: CA

WWE Wipo information: entry into national phase

Ref document number: 1994912849

Country of ref document: EP

WWP Wipo information: published in national office

Ref document number: 1994912849

Country of ref document: EP

WWW Wipo information: withdrawn in national office

Ref document number: 1994912849

Country of ref document: EP