US20030033477A1 - Method for raid striped I/O request generation using a shared scatter gather list - Google Patents
Method for raid striped I/O request generation using a shared scatter gather list Download PDFInfo
- Publication number
- US20030033477A1 US20030033477A1 US10/184,667 US18466702A US2003033477A1 US 20030033477 A1 US20030033477 A1 US 20030033477A1 US 18466702 A US18466702 A US 18466702A US 2003033477 A1 US2003033477 A1 US 2003033477A1
- Authority
- US
- United States
- Prior art keywords
- data
- request
- raid
- host
- requests
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/06—Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
- G06F3/0601—Interfaces specially adapted for storage systems
- G06F3/0602—Interfaces specially adapted for storage systems specifically adapted to achieve a particular effect
- G06F3/061—Improving I/O performance
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/06—Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
- G06F3/0601—Interfaces specially adapted for storage systems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/06—Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
- G06F3/0601—Interfaces specially adapted for storage systems
- G06F3/0628—Interfaces specially adapted for storage systems making use of a particular technique
- G06F3/0655—Vertical data movement, i.e. input-output transfer; data movement between one or more hosts and one or more storage devices
- G06F3/0659—Command handling arrangements, e.g. command buffers, queues, command scheduling
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/06—Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
- G06F3/0601—Interfaces specially adapted for storage systems
- G06F3/0668—Interfaces specially adapted for storage systems adopting a particular infrastructure
- G06F3/0671—In-line storage system
- G06F3/0683—Plurality of storage devices
- G06F3/0689—Disk arrays, e.g. RAID, JBOD
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/06—Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
- G06F2003/0697—Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers device management, e.g. handlers, drivers, I/O schedulers
Definitions
- the present invention generally relates to the field of information handling systems including computer systems and related devices using redundant array of independent disks (RAID) data storage systems and, more particularly, to a system and method for RAID striped data transfer.
- RAID redundant array of independent disks
- Striping is used in the implementation of most of the basic, single RAID levels and as a consequence, any multiple RAID levels that employ the single RAID levels.
- RAID 0 The simplest RAID level, uses block-level striping without parity. In RAID 0 , one file is divided into multiple groups of sectors and striped across multiple disks. Two advantages of RAID 0 include a large logical disk drive through drive spanning and performance acceleration through data striping. The major uses of RAID 0 are in situations where no redundancy is required or where redundancy can be provided through the use of transaction logs that recreate data from the last status recorded in the log. As a result, RAID 0 is often employed by users seeking high performance at low cost. However, the disadvantage of RAID 0 is that it provides no redundancy. If one disk fails, all data is lost and all disks must be reformatted. RAID 0 does not support a hot-spare drive, thus, if one drive fails, all data is lost. For comparison purposes, RAID 0 is regarded as the baseline against which to measure the performance of the other RAID levels.
- RAID enhanced also known as RAID 1 E, Hybrid RAID 1 , or RAID 6
- RAID 1 E shares the characteristics of RAID 1 but additionally allows more than two drives, including odd numbers of drives.
- the host operating system When the host operating system sends an I/O request to move data between the disk and its memory, it provides a scatter gather list (SGL) that describes the actual memory regions to move the data to or from.
- SGL scatter gather list
- the memory regions are fragmented and data is scattered to the various regions when data is transferred to host memory space.
- memory When memory is retrieved from host memory space and written to a drive, it is gathered from the various memory regions to form scatter gather list.
- striping is implemented by taking the Scatter Gather List (SGL) for the transfer data, which provides a description of the location of the data, and formulating a new SGL for the data transfer to each drive.
- SGL Scatter Gather List
- the original operating system host I/O request is divided into several new I/O requests to be sent to each of the stripe disk members and break up the original scatter gather list is divided into several smaller lists (one for each new I/O request). This method requires extra processing and memory resources to parse the original SGL and create new smaller ones dedicated for each of the new I/O requests.
- the present invention is directed to a more code efficient method for transferring data to and from disk drives.
- a single scatter gather list is generated according to parameters of the single I/O request.
- the single host I/O request is transmitted to and received by a disk array controller and is divided into a plurality of physical I/O requests in which each of the plurality of physical I/O requests traverses the same single scatter gather list.
- the method of the present invention is particularly useful for embedded RAID firmware for SCSI (Small Computer System Interface), FC (Fibre Channel), Serial Advanced Technology Attachment (SATA), Parallel Advanced Technology Attachment (PATA), and Serial Attached Small Computer System Interface (SAS) controllers.
- SCSI Small Computer System Interface
- FC Fibre Channel
- SATA Serial Advanced Technology Attachment
- PATA Parallel Advanced Technology Attachment
- SAS Serial Attached Small Computer System Interface
- FIG. 1 illustrates the method of the present invention
- FIG. 2 illustrates a system that employs the method of the present invention
- FIGS. 3A and 3B illustrate an SGL entry and SGL, respectively
- FIG. 4 illustrates an embodiment of the method for fragmented host memory
- FIG. 5 illustrates the method for generating a skip count
- FIG. 6 illustrates a RAID 0 implementation of the present invention
- FIG. 7 illustrates a RAID 1 E implementation of the present invention.
- the method of the present invention generates a single SGL for each host I/O request, processes the host I/O request into multiple physical I/O requests, shares the single SGL from the original request between all the physical I/O requests generated, and transfers data to and from multiple disks in a RAID system.
- the present invention may be practiced with a variety of Message Passing Technology (MPT) controllers, including Serial Attached Small Computer Systems Interface (SAS), Serial Advanced Technology Attachment (SATA), Parallel Advanced Technology Attachment (PATA), Fibre Channel (FC), Small Computer System Interface (SCSI), and the like.
- SAS Serial Attached Small Computer Systems Interface
- SATA Serial Advanced Technology Attachment
- PATA Parallel Advanced Technology Attachment
- FC Fibre Channel
- SCSI Small Computer System Interface
- the datapath engine is an essential part of all the MPT controllers.
- a datapath engine formed of dedicated hardware, software, firmware, or a combination traverses the SGL according to byte address offsets to access the block of bytes corresponding to a physical I/O request.
- the datapath engine skips ahead into the SGL an X number of bytes and then moves the bytes starting at that location.
- the skip ahead value is referred to as the SkipCount.
- This method may be designed for RAID 0 and RAID 1 B (enhanced mirroring) with various stripe sizes.
- the number of physical I/O requests generated is related to the number of stripes of data on or to be placed on the physical disks for a given host I/O request.
- FIG. 1 illustrates the method steps of the present invention.
- the host generates a single I/O request, per step 110 .
- the single host I/O request is transmitted to the data controller. Transmission may be conducted along a bus, through a switching device, by direct cabling, or in a wireless manner.
- the single host I/O request may be transmitted as parallel or serial data.
- the gathered data is assembled as a scatter gather list, as per step 120 .
- SGL the chain SGL
- each SGL entry contains an address and a length and may contain flags, such as Size of Address (i.e., 32-bit or 64-bit), End of List Reached, direction of data transfer, and the like.
- the SGL may be a set of pointers containing address locations which point to the stored bytes and information concerning the stored bytes, a data structure containing an initial address, address offsets, and length or start and stop addresses for each of the initial address and address offsets, and the like.
- the single host I/O request is received by the data controller and converted into multiple physical I/O requests, as per step 130 .
- the partition of the single host I/O request into multiple physical I/O requests may be accomplished through several parameters, such as Initial Stripe Size for indicating the size (i.e., the number of bytes) of the initial block of data (e.g., metadata or header information) to be transferred for the first stripe of the drive, an Initial Skip Size for indicating an amount of the drive (i.e., the number of bytes) to skip prior to transferring data for the drive, a Stripe Size for indicating the size (i.e., the number of bytes) in a single stripe, and a Stripe Skip for indicating the amount of the drive to skip after transferring a stripe of data.
- Initial Stripe Size for indicating the size (i.e., the number of bytes) of the initial block of data (e.g., metadata or header information) to be transferred for the first stripe of the drive
- an Initial Skip Size
- parameter information may included a set of absolute addresses, each having a start address and an end address, that determine the data that is to be transferred from host memory to the disk drives or determine where data is to be stored when transferred from the disk drives to host memory.
- a disk drive is selected.
- the order of access of the disk drives is predetermined.
- the order of access may follow a predetermined priority scheme. That is, the data may be written on a space available basis or may be written in a predetermined staggered order across the disk drives.
- Each physical I/O request corresponds to one data stripe.
- the data stripes may be of a generally uniform size or may consist of two or more different sizes.
- a data stripe may be 64K bytes, 1 MB, 0.5K byte (corresponding to a block of 512 bytes), or other appropriate size.
- step 140 the single scatter gather list remains intact during processing. Data for each physical I/O request is extracted as needed. The data extraction per physical I/O request may be performed through hardware, software, and/or firmware. The single scatter gather list may reside in host memory or may be copied to disk array controller memory.
- FIG. 2 illustrates an exemplary RAID system.
- the system 200 includes a controller, RAID controller 202 , for controlling the storage and retrieval of data between computer system 204 and one or more drives 206 , 208 , 210 , and 212 arranged in RAID array 214 .
- controller 202 may be implemented as either a separate hardware component, or alternately as software (i.e., sets of instructions), executed by a computer system such as computer system 204 .
- system 200 further includes a data path engine 216 for splitting data transferred from computer system 204 into blocks that are distributed across two or more drives 206 , 208 , 210 and 212 within the RAID array 214 .
- datapath engine 216 is shown as being implemented as part of controller 202 .
- datapath engine 216 may be implemented as sets of instructions (e.g., software, firmware, and the like) executed by controller 202 .
- datapath engine 216 may be implemented separately from controller 202 .
- datapath engine 216 may be implemented as a separate hardware component wherein controller 202 is implemented as hardware, or a separate software routine, wherein controller 202 is implemented via software.
- a portion of the datapath engine 216 may be implemented as firmware and a portion may be implemented dynamically through hardware or software for temporary storage of host and physical I/O request parameters.
- Datapath engine 216 utilizes the original SGL of the data to be transferred and parameters selected by the operator of the system 200 , for determining how data is to be split or striped across two or more of drives 206 , 208 , 210 , and 212 .
- the original SGL may be stored in the host memory and may be transferred to the disk array controller.
- FIGS. 3A and 3B illustrate an SGL list.
- an SGL entry 220 may contain a flag field.
- the flag field bits may be dedicated for certain parameter information, such as end of list, address size, direction read/write, local/system address, entry or element type (i.e., simple or chain), end of buffer, and last element.
- FIG. 3B shows an example of a scatter gather list 230 having five simple element SGL entries 220 that describe a total of 256K bytes.
- the present method for transferring stripes of data to and from a host device may be practiced in various ways. For example, several parameters or a single parameter may be used to traverse a single SGL.
- FIG. 4 illustrates another embodiment of the present invention.
- the data may be very fragmented in host memory.
- the resulting scatter gather list may be quite large to fully account for the data to be transferred through a host I/O request to accommodate the multiple start addresses and either byte lengths or end addresses which correspond to each data stripe.
- the SGL entry number n and stripe number m are initialized, per step 500 .
- Data is transferred according to the SGL entry number n and the stripe number m, per step 505 .
- a determination is made if all the data has been transferred which is referenced by SGL entry n, step 510 . If it has not been fully transferred, it is transferred to stripe m, per step 515 .
- step 515 may follow step 520 or steps 515 and 525 may be merged into a single transfer step. Data transfer may occur directly between the host device and the disk arrays or it may be temporarily stored within the disk array controller memory. Further, an initial determination may be made after step 500 as to whether there is actually data defined by the SGL, whether the disk drives are full, or whether there is enough space in a data stripe for the data. Also, step 505 may be merged into step 515 and/or step 525 .
- a single parameter, SkipCount is used to traverse the single SGL.
- FIG. 5 illustrates a method of determining the SkipCount.
- SkipCount is initialized, step 250 .
- a determination is made as to whether the host I/O request has been fulfilled, step 252 . This step may occur later in processing and may include other processing, such as determination of fault states including memory full. If the host I/O request has been fulfilled, the data transfer associated with the request is terminated, step 270 .
- a determination is made of the number of bytes m in associated with the SGL entry, step 254 . This may be achieved by being provided in the SGL, as well being determined by counting or other processing. The m bytes are transferred according to the SGL entry parameters, step 256 .
- the SkipCount is increased by m, for the number of bytes (or word) has been transferred for that SGL entry.
- the number of bytes may be limited to a small set of sizes, such as 64K, 63.5K, and 0.5K, or may be of a larger size to accommodate greater variation in the number of bytes associated with a given SGL entry.
- FIG. 6 illustrates an example of the present invention practiced with a RAID 0 in which a single parameter, SkipCount, is used to traverse the single SGL.
- the stripe size is 64K
- the number of drives is three
- the block size is 0.5K (512 bytes).
- the write I/O request is a 256K byte write starting at virtual logical block address (LBA) 1 .
- Five separate I/O requests are generated from the original I/O request to stripe the data across the three physical disks to accommodate four 64K stripes and a block stripe of 0.5K.
- a single physical I/O request is generated for each physical stripe that is being accessed.
- the datapath engine may transfer the data according to a data transfer scheme described above.
- I/O request A data stripe 0 is written to member disk 0 .
- I/O request B data stripe 1 is written to member disk 1 .
- I/O request C data stripe 2 is written to member disk 2 .
- I/O request D data stripe 3 is written to member disk 0 .
- I/O request E data stripe 4 is written to member disk 1 .
- Table 1 summarizes the mapping operation of data associated with the single host I/O request to the memory locations of the disk drives. Note that there is no redundancy or parity for RAID 0 .
- FIG. 7 illustrates a host I/O request for RAID 1 E in which a single parameter, SkipCount, is used to traverse the single SGL.
- the stripe size is 1 MB
- the number of drives is three
- the block size is 0.5K (512 bytes).
- a single host I/O request involves the transfer of 4 MB starting at virtual disk LBA 1 .
- Ten I/O requests are generated from the original I/O request to stripe the data across the three physical disks and to mirror it in the associated mirrored stripes.
- a single I/O request is generated for each physical stripe that is being accessed.
- the single I/O requests are as follows: in A, data stripe 0 is written to member disk 0 ; in B, data stripe 1 is written to member disk 1 ; and in C, data stripe 2 is written to member disk 2 . Furthermore, mirrored copies of these data stripes are written, in a staggered fashion to the disk drives.
- the mirrored I/O requests are as follows: in A′, stripe 0 ′ is written to member disk 1 ; in B′, stripes 1 ′ is written to member disk 2 ; and in C′ stripe 2 ′ is written to member disk 0 .
- single I/O requests D, E, F, D′, E′, and F′ are performed in a similar manner.
- Table 2 summarizes the mapping operation of data associated with the single host I/O request to the memory locations of the disk drives.
- I/O Request Size Drive Physical Stripe SkipCount A 1023.5 K 0 0 0 B 1024 K 1 1 1023.5 K C 1024 K 2 2 2047.5 K D 1024 K 0 3 3071.5 K E 0.5 K 1 4 4095.5 K A' 1023.5 K 1 0' 0 B' 1024 K 2 1' 1023.5 K C' 1024 K 0 2' 2047.5 K D' 1024 K 1 3' 3071.5 K E' 0.5 K 2 4' 4095.5 K
- the present invention is not limited to the embodiments described above. Although the present invention has disclosed data stripe sizes of 64K, 1 MB, and 0.5K, it may be practiced with other data sizes. Generally, stripe size is a power of 2. The data stripe sizes may be 128K, 256K, 512K, or the like. Also, the present invention is not limited to RAID 0 and RAID 1 E systems, but may be practiced with other RAID levels, such as RAID 1 and RAID 10 , as well as other disk array technology. Each physical I/O request may be created for an entire stripe or a part of a stripe.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
Description
- The present invention is a continuation-in-part of U.S. patent application Ser. No. 09/795,825, entitled “Method and Apparatus for RAID Striped Data Transfer”, filed Feb. 28, 2001, herein incorporated by reference.
- The present invention generally relates to the field of information handling systems including computer systems and related devices using redundant array of independent disks (RAID) data storage systems and, more particularly, to a system and method for RAID striped data transfer.
- The performance of disk storage systems is limited by the speed of mechanical components that are used for positioning and transferring data. Since RAID storage systems utilize multiple drives, it is possible to improve their performance by using the hardware of all drives in parallel. Drive spanning allows multiple physical disk drives to be logically concatenated into a single logical disk drive. The capacity of the logical drive created via spanning is the capacity of the physical drives times the number of physical drives. A technique called data striping distributes data evenly across the physical drives in such a manner as to maximize input/output (I/O) performance. Striping divides the logical drive into data blocks called stripes, which are then distributed over the physical disk drives. The layout is such that a sequential read of data on the logical drive results in parallel reads to each of the physical drives. This results in improved performance since multiple drives are operating simultaneously.
- Striping is used in the implementation of most of the basic, single RAID levels and as a consequence, any multiple RAID levels that employ the single RAID levels.
- The simplest RAID level,
RAID 0, uses block-level striping without parity. InRAID 0, one file is divided into multiple groups of sectors and striped across multiple disks. Two advantages ofRAID 0 include a large logical disk drive through drive spanning and performance acceleration through data striping. The major uses ofRAID 0 are in situations where no redundancy is required or where redundancy can be provided through the use of transaction logs that recreate data from the last status recorded in the log. As a result,RAID 0 is often employed by users seeking high performance at low cost. However, the disadvantage ofRAID 0 is that it provides no redundancy. If one disk fails, all data is lost and all disks must be reformatted.RAID 0 does not support a hot-spare drive, thus, if one drive fails, all data is lost. For comparison purposes,RAID 0 is regarded as the baseline against which to measure the performance of the other RAID levels. - Another RAID level, RAID enhanced (also known as RAID1E,
Hybrid RAID 1, or RAID 6), combines mirroring with data striping—data is striped across each disk in the array. The first set of stripes includes the data stripes, and the second set of stripes is the mirror (copies) of the first data stripe, but shifted one drive. RAID 1E shares the characteristics ofRAID 1 but additionally allows more than two drives, including odd numbers of drives. - When the host operating system sends an I/O request to move data between the disk and its memory, it provides a scatter gather list (SGL) that describes the actual memory regions to move the data to or from. Typically, the memory regions are fragmented and data is scattered to the various regions when data is transferred to host memory space. When memory is retrieved from host memory space and written to a drive, it is gathered from the various memory regions to form scatter gather list. Typically, striping is implemented by taking the Scatter Gather List (SGL) for the transfer data, which provides a description of the location of the data, and formulating a new SGL for the data transfer to each drive. The original operating system host I/O request is divided into several new I/O requests to be sent to each of the stripe disk members and break up the original scatter gather list is divided into several smaller lists (one for each new I/O request). This method requires extra processing and memory resources to parse the original SGL and create new smaller ones dedicated for each of the new I/O requests.
- Therefore, it would be desirable to provide a code efficient method of fulfilling a host I/O request that does not require extra processing and extra memory resources.
- Accordingly, the present invention is directed to a more code efficient method for transferring data to and from disk drives.
- In a method for processing a single I/O request from a host to a disk array controller, a single scatter gather list is generated according to parameters of the single I/O request. The single host I/O request is transmitted to and received by a disk array controller and is divided into a plurality of physical I/O requests in which each of the plurality of physical I/O requests traverses the same single scatter gather list. By generating a physical I/O request for each physical stripe and sharing the same SGL by the physical I/O requests, processor action is reduced (i.e., fewer physical I/O requests are generated per host I/O request) resulting in lower costs. Drives are also very efficient at storing I/O requests.
- The method of the present invention is particularly useful for embedded RAID firmware for SCSI (Small Computer System Interface), FC (Fibre Channel), Serial Advanced Technology Attachment (SATA), Parallel Advanced Technology Attachment (PATA), and Serial Attached Small Computer System Interface (SAS) controllers.
- It is to be understood that both the forgoing general description and the following detailed description are exemplary and explanatory only and are not restrictive of the invention as claimed. The accompanying drawings, which are incorporated in and constitute a part of the specification, illustrate an embodiment of the invention and together with the general description, serve to explain the principles of the invention.
- The numerous advantages of the present invention may be better understood by those skilled in the art by reference to the accompanying figures in which:
- FIG. 1 illustrates the method of the present invention;
- FIG. 2 illustrates a system that employs the method of the present invention;
- FIGS. 3A and 3B illustrate an SGL entry and SGL, respectively;
- FIG. 4 illustrates an embodiment of the method for fragmented host memory;
- FIG. 5 illustrates the method for generating a skip count;
- FIG. 6 illustrates a
RAID 0 implementation of the present invention; and - FIG. 7 illustrates a RAID1E implementation of the present invention.
- Reference will now be made in detail to the presently preferred embodiments of the invention, examples of which are illustrated in the accompanying drawings.
- The method of the present invention generates a single SGL for each host I/O request, processes the host I/O request into multiple physical I/O requests, shares the single SGL from the original request between all the physical I/O requests generated, and transfers data to and from multiple disks in a RAID system. The present invention may be practiced with a variety of Message Passing Technology (MPT) controllers, including Serial Attached Small Computer Systems Interface (SAS), Serial Advanced Technology Attachment (SATA), Parallel Advanced Technology Attachment (PATA), Fibre Channel (FC), Small Computer System Interface (SCSI), and the like. The datapath engine is an essential part of all the MPT controllers. A datapath engine formed of dedicated hardware, software, firmware, or a combination traverses the SGL according to byte address offsets to access the block of bytes corresponding to a physical I/O request. In effect, the datapath engine skips ahead into the SGL an X number of bytes and then moves the bytes starting at that location. The skip ahead value is referred to as the SkipCount. This method may be designed for
RAID 0 and RAID 1B (enhanced mirroring) with various stripe sizes. The number of physical I/O requests generated is related to the number of stripes of data on or to be placed on the physical disks for a given host I/O request. - FIG. 1 illustrates the method steps of the present invention. The host generates a single I/O request, per
step 110. The single host I/O request is transmitted to the data controller. Transmission may be conducted along a bus, through a switching device, by direct cabling, or in a wireless manner. The single host I/O request may be transmitted as parallel or serial data. - The gathered data is assembled as a scatter gather list, as per
step 120. One of the types of SGL, the chain SGL, may be useful where the data is located in noncontiguous physical memory. In general, each SGL entry contains an address and a length and may contain flags, such as Size of Address (i.e., 32-bit or 64-bit), End of List Reached, direction of data transfer, and the like. The SGL may be a set of pointers containing address locations which point to the stored bytes and information concerning the stored bytes, a data structure containing an initial address, address offsets, and length or start and stop addresses for each of the initial address and address offsets, and the like. - The single host I/O request is received by the data controller and converted into multiple physical I/O requests, as per
step 130. The partition of the single host I/O request into multiple physical I/O requests may be accomplished through several parameters, such as Initial Stripe Size for indicating the size (i.e., the number of bytes) of the initial block of data (e.g., metadata or header information) to be transferred for the first stripe of the drive, an Initial Skip Size for indicating an amount of the drive (i.e., the number of bytes) to skip prior to transferring data for the drive, a Stripe Size for indicating the size (i.e., the number of bytes) in a single stripe, and a Stripe Skip for indicating the amount of the drive to skip after transferring a stripe of data. Alternatively, parameter information may included a set of absolute addresses, each having a start address and an end address, that determine the data that is to be transferred from host memory to the disk drives or determine where data is to be stored when transferred from the disk drives to host memory. As part of the step of generating multiple physical I/O requests, a disk drive is selected. In the case where data is being read from the disk drives, the order of access of the disk drives is predetermined. In the case where data is being written to the disk drives, the order of access may follow a predetermined priority scheme. That is, the data may be written on a space available basis or may be written in a predetermined staggered order across the disk drives. Each physical I/O request corresponds to one data stripe. The data stripes may be of a generally uniform size or may consist of two or more different sizes. A data stripe may be 64K bytes, 1 MB, 0.5K byte (corresponding to a block of 512 bytes), or other appropriate size. - In
step 140, the single scatter gather list remains intact during processing. Data for each physical I/O request is extracted as needed. The data extraction per physical I/O request may be performed through hardware, software, and/or firmware. The single scatter gather list may reside in host memory or may be copied to disk array controller memory. - FIG. 2 illustrates an exemplary RAID system. The
system 200 includes a controller,RAID controller 202, for controlling the storage and retrieval of data betweencomputer system 204 and one ormore drives RAID array 214. In embodiments of the invention,controller 202 may be implemented as either a separate hardware component, or alternately as software (i.e., sets of instructions), executed by a computer system such ascomputer system 204. - In accordance with the present invention,
system 200 further includes adata path engine 216 for splitting data transferred fromcomputer system 204 into blocks that are distributed across two ormore drives RAID array 214. In the exemplary embodiment shown in FIG. 2,datapath engine 216 is shown as being implemented as part ofcontroller 202. For instance, in one embodiment of the invention,datapath engine 216 may be implemented as sets of instructions (e.g., software, firmware, and the like) executed bycontroller 202. Alternately,datapath engine 216 may be implemented separately fromcontroller 202. For instance,datapath engine 216 may be implemented as a separate hardware component whereincontroller 202 is implemented as hardware, or a separate software routine, whereincontroller 202 is implemented via software. A portion of thedatapath engine 216 may be implemented as firmware and a portion may be implemented dynamically through hardware or software for temporary storage of host and physical I/O request parameters. -
Datapath engine 216 utilizes the original SGL of the data to be transferred and parameters selected by the operator of thesystem 200, for determining how data is to be split or striped across two or more ofdrives - FIGS. 3A and 3B illustrate an SGL list. In FIG. 3A, an
SGL entry 220 may contain a flag field. The flag field bits may be dedicated for certain parameter information, such as end of list, address size, direction read/write, local/system address, entry or element type (i.e., simple or chain), end of buffer, and last element. FIG. 3B shows an example of a scatter gatherlist 230 having five simpleelement SGL entries 220 that describe a total of 256K bytes. - The present method for transferring stripes of data to and from a host device may be practiced in various ways. For example, several parameters or a single parameter may be used to traverse a single SGL.
- FIG. 4 illustrates another embodiment of the present invention. In this embodiment, the data may be very fragmented in host memory. The resulting scatter gather list may be quite large to fully account for the data to be transferred through a host I/O request to accommodate the multiple start addresses and either byte lengths or end addresses which correspond to each data stripe. In the method, the SGL entry number n and stripe number m are initialized, per
step 500. Data is transferred according to the SGL entry number n and the stripe number m, perstep 505. A determination is made if all the data has been transferred which is referenced by SGL entry n,step 510. If it has not been fully transferred, it is transferred to stripe m, perstep 515. Otherwise, a determination is made as to whether the end of the SGL has been reached, perstep 530. If the end of the SGL has not been reached and the data defined by the last entry of the SGL has been fully transferred, the next entry in the SGL is accessed, perstep 540. Otherwise, transfer for the single host I/O request is ended, perstep 535. If data is transferred as defined by SGL entry n, as perstep 515, a determination is then made as to whether the end of stripe m has been reached, perstep 520. If it has, then another determination is made whether this is the last stripe,step 550. If it is, then transfer for the host I/O request ends, perstep 555. Otherwise, the next stripe m is accessed, perstep 545. The data is then written to or read from stripe m, perstep 525, and processing returns to step 510. Variations of this method are contemplated by the present invention. For example, step 515 may followstep 520 orsteps step 500 as to whether there is actually data defined by the SGL, whether the disk drives are full, or whether there is enough space in a data stripe for the data. Also, step 505 may be merged intostep 515 and/or step 525. - In another embodiment, a single parameter, SkipCount, is used to traverse the single SGL. FIG. 5 illustrates a method of determining the SkipCount. SkipCount is initialized,
step 250. A determination is made as to whether the host I/O request has been fulfilled,step 252. This step may occur later in processing and may include other processing, such as determination of fault states including memory full. If the host I/O request has been fulfilled, the data transfer associated with the request is terminated,step 270. A determination is made of the number of bytes m in associated with the SGL entry,step 254. This may be achieved by being provided in the SGL, as well being determined by counting or other processing. The m bytes are transferred according to the SGL entry parameters,step 256. After the transfer for the SGL entry is finished, the SkipCount is increased by m, for the number of bytes (or word) has been transferred for that SGL entry. The number of bytes may be limited to a small set of sizes, such as 64K, 63.5K, and 0.5K, or may be of a larger size to accommodate greater variation in the number of bytes associated with a given SGL entry. - FIG. 6 illustrates an example of the present invention practiced with a
RAID 0 in which a single parameter, SkipCount, is used to traverse the single SGL. In this example, the stripe size is 64K, the number of drives is three, and the block size is 0.5K (512 bytes). The write I/O request is a 256K byte write starting at virtual logical block address (LBA) 1. Five separate I/O requests are generated from the original I/O request to stripe the data across the three physical disks to accommodate four 64K stripes and a block stripe of 0.5K. A single physical I/O request is generated for each physical stripe that is being accessed. The datapath engine may transfer the data according to a data transfer scheme described above. During I/O request A,data stripe 0 is written tomember disk 0. During I/O request B,data stripe 1 is written tomember disk 1. During I/O request C,data stripe 2 is written tomember disk 2. During I/O request D,data stripe 3 is written tomember disk 0. During I/O request E,data stripe 4 is written tomember disk 1. Table 1 summarizes the mapping operation of data associated with the single host I/O request to the memory locations of the disk drives. Note that there is no redundancy or parity forRAID 0.TABLE 1 I/O Starting Starting request Size Drive Virtual LBA Physical LBA SkipCount A 63.5 K 0 1 1 0 B 64 K 1 128 0 63.5 K C 64 K 2 256 0 127.5 K D 64 K 0 384 128 191.5 K E 0.5 K 1 512 128 255.5 K - FIG. 7 illustrates a host I/O request for RAID1E in which a single parameter, SkipCount, is used to traverse the single SGL. In this example, the stripe size is 1 MB, the number of drives is three, and the block size is 0.5K (512 bytes). A single host I/O request involves the transfer of 4 MB starting at
virtual disk LBA 1. Ten I/O requests are generated from the original I/O request to stripe the data across the three physical disks and to mirror it in the associated mirrored stripes. A single I/O request is generated for each physical stripe that is being accessed. The single I/O requests are as follows: in A,data stripe 0 is written tomember disk 0; in B,data stripe 1 is written tomember disk 1; and in C,data stripe 2 is written tomember disk 2. Furthermore, mirrored copies of these data stripes are written, in a staggered fashion to the disk drives. The mirrored I/O requests are as follows: in A′,stripe 0′ is written tomember disk 1; in B′,stripes 1′ is written tomember disk 2; and in C′stripe 2′ is written tomember disk 0. Next, single I/O requests D, E, F, D′, E′, and F′ are performed in a similar manner. Table 2 summarizes the mapping operation of data associated with the single host I/O request to the memory locations of the disk drives.TABLE 2 I/O Request Size Drive Physical Stripe SkipCount A 1023.5 K 0 0 0 B 1024 K 1 1 1023.5 K C 1024 K 2 2 2047.5 K D 1024 K 0 3 3071.5 K E 0.5 K 1 4 4095.5 K A' 1023.5 K 1 0' 0 B' 1024 K 2 1' 1023.5 K C' 1024 K 0 2' 2047.5 K D' 1024 K 1 3' 3071.5 K E' 0.5 K 2 4' 4095.5 K - The present invention is not limited to the embodiments described above. Although the present invention has disclosed data stripe sizes of 64K, 1 MB, and 0.5K, it may be practiced with other data sizes. Generally, stripe size is a power of 2. The data stripe sizes may be 128K, 256K, 512K, or the like. Also, the present invention is not limited to
RAID 0 and RAID 1E systems, but may be practiced with other RAID levels, such asRAID 1 and RAID 10, as well as other disk array technology. Each physical I/O request may be created for an entire stripe or a part of a stripe. - It is believed that the present invention and many of its attendant advantages will be understood by the forgoing description. It is also believed that it will be apparent that various changes may be made in the form, construction and arrangement of the components thereof without departing from the scope and spirit of the invention or without sacrificing all of its material advantages. The form herein before described being merely an explanatory embodiment thereof. It is the intention of the following claims to encompass and include such changes.
Claims (20)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US10/184,667 US7155569B2 (en) | 2001-02-28 | 2002-06-28 | Method for raid striped I/O request generation using a shared scatter gather list |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US79582501A | 2001-02-28 | 2001-02-28 | |
US10/184,667 US7155569B2 (en) | 2001-02-28 | 2002-06-28 | Method for raid striped I/O request generation using a shared scatter gather list |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US79582501A Continuation-In-Part | 2001-02-28 | 2001-02-28 |
Publications (2)
Publication Number | Publication Date |
---|---|
US20030033477A1 true US20030033477A1 (en) | 2003-02-13 |
US7155569B2 US7155569B2 (en) | 2006-12-26 |
Family
ID=25166553
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/184,667 Expired - Fee Related US7155569B2 (en) | 2001-02-28 | 2002-06-28 | Method for raid striped I/O request generation using a shared scatter gather list |
Country Status (1)
Country | Link |
---|---|
US (1) | US7155569B2 (en) |
Cited By (71)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030212860A1 (en) * | 2002-03-01 | 2003-11-13 | Hitachi, Ltd. | System and method for storage system |
US20040162926A1 (en) * | 2003-02-14 | 2004-08-19 | Itzhak Levy | Serial advanced technology attachment interface |
WO2004104846A2 (en) * | 2003-05-26 | 2004-12-02 | Koninklijke Philips Electronics N.V. | Method and device for transferring data between a main memory and a storage device |
US20050005063A1 (en) * | 2003-07-02 | 2005-01-06 | Ling-Yi Liu | Jbod subsystem and external emulation controller thereof |
US20050138154A1 (en) * | 2003-12-18 | 2005-06-23 | Intel Corporation | Enclosure management device |
US20050138191A1 (en) * | 2003-12-18 | 2005-06-23 | Pak-Lung Seto | Adaptor supporting different protocols |
US7062501B1 (en) * | 2001-08-08 | 2006-06-13 | Adaptec, Inc. | Structure and method for linking scatter/gather list segments for host adapters |
US20060136666A1 (en) * | 2004-12-21 | 2006-06-22 | Ching-Te Pang | SAS storage virtualization controller, subsystem and system using the same, and method therefor |
US20060190689A1 (en) * | 2003-03-25 | 2006-08-24 | Koninklijke Philips Electronics N.V. | Method of addressing data in a shared memory by means of an offset |
US20060190763A1 (en) * | 2005-02-24 | 2006-08-24 | Dot Hill Systems Corp. | Redundant storage array method and apparatus |
WO2007012919A2 (en) * | 2005-07-27 | 2007-02-01 | Adaptec, Inc. | Ripple queuing algorithm for a sas wide-port raid controller |
US7188212B2 (en) * | 2004-05-06 | 2007-03-06 | International Business Machines Corporation | Method and system for storing data in an array of storage devices with additional and autonomic protection |
US20070067537A1 (en) * | 2003-12-18 | 2007-03-22 | Pak-Lung Seto | Multiple interfaces in a storage enclosure |
US20070106869A1 (en) * | 2005-11-04 | 2007-05-10 | Sun Microsystems, Inc. | Method and system for dirty time logging |
US20070106677A1 (en) * | 2005-11-04 | 2007-05-10 | Sun Microsystems, Inc. | Method and system for pruned resilvering using a dirty time log |
US20070106866A1 (en) * | 2005-11-04 | 2007-05-10 | Sun Microsystems, Inc. | Method and system for metadata-based resilvering |
US20070106863A1 (en) * | 2005-11-04 | 2007-05-10 | Sun Microsystems, Inc. | Method and system for storing a sparse file using fill counts |
US20070106870A1 (en) * | 2005-11-04 | 2007-05-10 | Sun Microsystems, Inc. | Method and system for block reallocation |
US20070106865A1 (en) * | 2005-11-04 | 2007-05-10 | Sun Microsystems, Inc. | Method and system for using a block allocation policy |
US20070106867A1 (en) * | 2005-11-04 | 2007-05-10 | Sun Microsystems, Inc. | Method and system for dirty time log directed resilvering |
US20070106706A1 (en) * | 2005-11-04 | 2007-05-10 | Sun Microsystems, Inc. | Unlimited file system snapshots and clones |
US20070106868A1 (en) * | 2005-11-04 | 2007-05-10 | Sun Microsystems, Inc. | Method and system for latency-directed block allocation |
US20070106864A1 (en) * | 2005-11-04 | 2007-05-10 | Sun Microsystems, Inc. | Multiple replication levels with pooled devices |
US20070106925A1 (en) * | 2005-11-04 | 2007-05-10 | Sun Microsystems, Inc. | Method and system using checksums to repair data |
US20070106851A1 (en) * | 2005-11-04 | 2007-05-10 | Sun Microsystems, Inc. | Method and system supporting per-file and per-block replication |
US20070106862A1 (en) * | 2005-11-04 | 2007-05-10 | Sun Microsystems, Inc. | Ditto blocks |
US20070106632A1 (en) * | 2005-11-04 | 2007-05-10 | Sun Microsystems, Inc. | Method and system for object allocation using fill counts |
US20070113024A1 (en) * | 2003-12-22 | 2007-05-17 | Kabushiki Kaisha Kawai Gakki | Device for processing access concurrence to shared memory |
US20070112895A1 (en) * | 2005-11-04 | 2007-05-17 | Sun Microsystems, Inc. | Block-based incremental backup |
US20070118576A1 (en) * | 2005-11-04 | 2007-05-24 | Sun Microsystems, Inc. | Method and system for adaptive metadata replication |
US20070124659A1 (en) * | 2005-11-04 | 2007-05-31 | Sun Microsystems, Inc. | Method and system for data replication |
US20070133536A1 (en) * | 2005-12-08 | 2007-06-14 | International Business Machines Corporation | Method and apparatus for striping message payload data over a network |
US20070174494A1 (en) * | 2005-11-04 | 2007-07-26 | Sun Microsystems, Inc. | I/O dependency graphs |
US20070198889A1 (en) * | 2005-11-04 | 2007-08-23 | Sun Microsystems, Inc. | Method and system for repairing partially damaged blocks |
US7281188B1 (en) | 2004-05-26 | 2007-10-09 | Sun Microsystems, Inc. | Method and system for detecting and correcting data errors using data permutations |
US20080104355A1 (en) * | 2006-10-31 | 2008-05-01 | Sun Microsystems, Inc. | Method and system for reallocating blocks in a storage pool |
US20080104150A1 (en) * | 2006-10-31 | 2008-05-01 | Sun Microsystems, Inc. | Method and system for priority-based allocation in a storage pool |
US20080104204A1 (en) * | 2006-10-31 | 2008-05-01 | Sun Microsystems, Inc. | Method and apparatus for power-managing storage devices in a storage pool |
US7412450B1 (en) | 2004-05-26 | 2008-08-12 | Sun Microsystems, Inc. | Method and apparatus for identifying tampering of data in a file system |
US7415653B1 (en) | 2004-04-21 | 2008-08-19 | Sun Microsystems, Inc. | Method and apparatus for vectored block-level checksum for file system data integrity |
US7424574B1 (en) * | 2004-04-21 | 2008-09-09 | Sun Microsystems, Inc. | Method and apparatus for dynamic striping |
US7437528B1 (en) | 2004-08-17 | 2008-10-14 | Sun Microsystems, Inc. | Gang blocks |
US7496586B1 (en) | 2004-05-26 | 2009-02-24 | Sun Microsystems, Inc. | Method and apparatus for compressing data in a file system |
US20090089343A1 (en) * | 2007-09-27 | 2009-04-02 | Sun Microsystems, Inc. | Method and system for block allocation for hybrid drives |
US7526622B1 (en) | 2004-05-26 | 2009-04-28 | Sun Microsystems, Inc. | Method and system for detecting and correcting data errors using checksums and replication |
US7533225B1 (en) | 2004-08-17 | 2009-05-12 | Sun Microsystems, Inc. | Method and apparatus for enabling adaptive endianness |
US20090198885A1 (en) * | 2008-02-04 | 2009-08-06 | Manoj Jose K | System and methods for host software stripe management in a striped storage subsystem |
US20090235278A1 (en) * | 2008-03-14 | 2009-09-17 | Prakash Babu H | Method for tracking and/or verifying message passing in a simulation environment |
US7603568B1 (en) | 2004-04-21 | 2009-10-13 | Sun Microsystems, Inc. | Method and apparatus for self-validating checksums in a file system |
US20090265519A1 (en) * | 2008-04-18 | 2009-10-22 | Sun Microsystems, Inc. | Method and system for power aware i/o scheduling |
US20090313446A1 (en) * | 2008-06-12 | 2009-12-17 | Sun Microsystems, Inc. | Method and system for cross-domain data sharing |
US20090327602A1 (en) * | 2008-06-30 | 2009-12-31 | Sun Microsystems, Inc. | Method and system for managing wear-level aware file systems |
US20100332446A1 (en) * | 2009-06-29 | 2010-12-30 | Sun Microsystems, Inc. | Storage pool scrubbing with concurrent snapshots |
US8645623B1 (en) * | 2007-06-28 | 2014-02-04 | Emc Corporation | Method for performing a raid operation in a data storage system |
CN103617136A (en) * | 2013-12-04 | 2014-03-05 | 华为技术有限公司 | SCSI drive side and I/O request control method |
US20140244938A1 (en) * | 2013-02-27 | 2014-08-28 | Vmware, Inc. | Method and Apparatus for Returning Reads in the Presence of Partial Data Unavailability |
US20140304441A1 (en) * | 2013-04-09 | 2014-10-09 | Apple Inc. | Protocol conversion involving multiple virtual channels |
US20170048320A1 (en) * | 2015-08-13 | 2017-02-16 | Advanced Micro Devices, Inc. | Distributed gather/scatter operations across a network of memory nodes |
US9778858B1 (en) * | 2015-02-11 | 2017-10-03 | Microsemi Solutions (U.S.), Inc. | Apparatus and method for scatter gather list handling for an out of order system |
CN108763023A (en) * | 2018-05-29 | 2018-11-06 | 郑州云海信息技术有限公司 | A kind of stage division of disk, device, equipment and readable storage medium storing program for executing |
US20190073141A1 (en) * | 2017-09-07 | 2019-03-07 | Pure Storage, Inc. | Converting raid data between persistent storage types |
US10522199B2 (en) * | 2015-02-06 | 2019-12-31 | Micron Technology, Inc. | Apparatuses and methods for scatter and gather |
US10579282B1 (en) * | 2016-03-30 | 2020-03-03 | EMC IP Holding Company LLC | Distributed copy in multi-copy replication where offset and size of I/O requests to replication site is half offset and size of I/O request to production volume |
CN112988623A (en) * | 2019-12-17 | 2021-06-18 | 北京忆芯科技有限公司 | Method and storage device for accelerating SGL (secure gateway) processing |
US11083381B2 (en) | 2009-09-11 | 2021-08-10 | University Of Virginia Patent Foundation | Systems and methods for determining pressure frequency changes in a subject |
US11093180B2 (en) * | 2019-09-27 | 2021-08-17 | Dell Products L.P. | RAID storage multi-operation command system |
CN113742258A (en) * | 2020-05-29 | 2021-12-03 | 慧与发展有限责任合伙企业 | System and method for efficient cache coherency protocol processing |
US11593036B2 (en) | 2017-06-12 | 2023-02-28 | Pure Storage, Inc. | Staging data within a unified storage element |
US11609718B1 (en) | 2017-06-12 | 2023-03-21 | Pure Storage, Inc. | Identifying valid data after a storage system recovery |
US11775430B1 (en) * | 2018-03-12 | 2023-10-03 | Amazon Technologies, Inc. | Memory access for multiple circuit components |
US11960777B2 (en) | 2017-06-12 | 2024-04-16 | Pure Storage, Inc. | Utilizing multiple redundancy schemes within a unified storage element |
Families Citing this family (28)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7761630B2 (en) * | 2005-06-07 | 2010-07-20 | Lsi Corporation | Application programming interface for fusion message passing technology |
US8060696B2 (en) * | 2007-04-27 | 2011-11-15 | Siemens Medical Solutions Usa, Inc. | Positron emission tomography event stream buffering |
US8478915B2 (en) | 2008-02-14 | 2013-07-02 | International Business Machines Corporation | Determining extended capability of a channel path |
US7941570B2 (en) | 2008-02-14 | 2011-05-10 | International Business Machines Corporation | Bi-directional data transfer within a single I/O operation |
US8312189B2 (en) | 2008-02-14 | 2012-11-13 | International Business Machines Corporation | Processing of data to monitor input/output operations |
US9052837B2 (en) | 2008-02-14 | 2015-06-09 | International Business Machines Corporation | Processing communication data in a ships passing condition |
US7890668B2 (en) | 2008-02-14 | 2011-02-15 | International Business Machines Corporation | Providing indirect data addressing in an input/output processing system where the indirect data address list is non-contiguous |
US8117347B2 (en) | 2008-02-14 | 2012-02-14 | International Business Machines Corporation | Providing indirect data addressing for a control block at a channel subsystem of an I/O processing system |
US8214562B2 (en) | 2008-02-14 | 2012-07-03 | International Business Machines Corporation | Processing of data to perform system changes in an input/output processing system |
US8332542B2 (en) | 2009-11-12 | 2012-12-11 | International Business Machines Corporation | Communication with input/output system devices |
US8402216B1 (en) * | 2010-03-17 | 2013-03-19 | Symantec Corporation | Systems and methods for off-host backups |
US8230189B1 (en) * | 2010-03-17 | 2012-07-24 | Symantec Corporation | Systems and methods for off-host backups of striped volumes |
US9021155B2 (en) | 2011-06-01 | 2015-04-28 | International Business Machines Corporation | Fibre channel input/output data routing including discarding of data transfer requests in response to error detection |
US8583988B2 (en) | 2011-06-01 | 2013-11-12 | International Business Machines Corporation | Fibre channel input/output data routing system and method |
US8738811B2 (en) | 2011-06-01 | 2014-05-27 | International Business Machines Corporation | Fibre channel input/output data routing system and method |
US8677027B2 (en) | 2011-06-01 | 2014-03-18 | International Business Machines Corporation | Fibre channel input/output data routing system and method |
US8364853B2 (en) | 2011-06-01 | 2013-01-29 | International Business Machines Corporation | Fibre channel input/output data routing system and method |
US8364854B2 (en) | 2011-06-01 | 2013-01-29 | International Business Machines Corporation | Fibre channel input/output data routing system and method |
US8819302B2 (en) | 2011-06-13 | 2014-08-26 | Lsi Corporation | System to manage input/output performance and/or deadlock in network attached storage gateway connected to a storage area network environment |
US8549185B2 (en) | 2011-06-30 | 2013-10-01 | International Business Machines Corporation | Facilitating transport mode input/output operations between a channel subsystem and input/output devices |
US8346978B1 (en) | 2011-06-30 | 2013-01-01 | International Business Machines Corporation | Facilitating transport mode input/output operations between a channel subsystem and input/output devices |
US8473641B2 (en) | 2011-06-30 | 2013-06-25 | International Business Machines Corporation | Facilitating transport mode input/output operations between a channel subsystem and input/output devices |
US8312176B1 (en) | 2011-06-30 | 2012-11-13 | International Business Machines Corporation | Facilitating transport mode input/output operations between a channel subsystem and input/output devices |
US8918542B2 (en) | 2013-03-15 | 2014-12-23 | International Business Machines Corporation | Facilitating transport mode data transfer between a channel subsystem and input/output devices |
TW201445429A (en) | 2013-05-22 | 2014-12-01 | Asmedia Technology Inc | Disk array system and data processing method |
US8990439B2 (en) | 2013-05-29 | 2015-03-24 | International Business Machines Corporation | Transport mode data transfer between a channel subsystem and input/output devices |
US10031685B2 (en) * | 2015-05-14 | 2018-07-24 | Quanta Computer Inc. | Power management of storage subsystem |
US10282301B2 (en) | 2017-05-18 | 2019-05-07 | Avago Technologies International Sales Pte. Limited | Method and system for hardware accelerated read-ahead caching |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6591339B1 (en) * | 1999-05-03 | 2003-07-08 | 3Ware, Inc. | Methods and systems for selecting block sizes for use with disk arrays |
US6591310B1 (en) * | 2000-05-11 | 2003-07-08 | Lsi Logic Corporation | Method of responding to I/O request and associated reply descriptor |
US6834326B1 (en) * | 2000-02-04 | 2004-12-21 | 3Com Corporation | RAID method and device with network protocol between controller and storage devices |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5483641A (en) * | 1991-12-17 | 1996-01-09 | Dell Usa, L.P. | System for scheduling readahead operations if new request is within a proximity of N last read requests wherein N is dependent on independent activities |
US5909516A (en) * | 1996-03-29 | 1999-06-01 | Sarnoff Corporation | Method and apparatus for decomposing an image stream into units of local contrast |
US6021462A (en) * | 1997-08-29 | 2000-02-01 | Apple Computer, Inc. | Methods and apparatus for system memory efficient disk access to a raid system using stripe control information |
US6381674B2 (en) * | 1997-09-30 | 2002-04-30 | Lsi Logic Corporation | Method and apparatus for providing centralized intelligent cache between multiple data controlling elements |
-
2002
- 2002-06-28 US US10/184,667 patent/US7155569B2/en not_active Expired - Fee Related
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6591339B1 (en) * | 1999-05-03 | 2003-07-08 | 3Ware, Inc. | Methods and systems for selecting block sizes for use with disk arrays |
US6834326B1 (en) * | 2000-02-04 | 2004-12-21 | 3Com Corporation | RAID method and device with network protocol between controller and storage devices |
US6591310B1 (en) * | 2000-05-11 | 2003-07-08 | Lsi Logic Corporation | Method of responding to I/O request and associated reply descriptor |
Cited By (130)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7062501B1 (en) * | 2001-08-08 | 2006-06-13 | Adaptec, Inc. | Structure and method for linking scatter/gather list segments for host adapters |
US7181571B2 (en) | 2002-03-01 | 2007-02-20 | Hitachi, Ltd. | System and method for storage system |
US20030212860A1 (en) * | 2002-03-01 | 2003-11-13 | Hitachi, Ltd. | System and method for storage system |
US7596660B2 (en) | 2002-03-01 | 2009-09-29 | Hitachi, Ltd. | System and method for storage system |
US20070136527A1 (en) * | 2002-03-01 | 2007-06-14 | Hitachi, Ltd. | System and method for storage system |
US6985997B2 (en) | 2002-03-01 | 2006-01-10 | Hitachi, Ltd. | System and method for storage system |
US20060020755A1 (en) * | 2002-03-01 | 2006-01-26 | Hitachi, Ltd. | System and method for storage system |
US20040162926A1 (en) * | 2003-02-14 | 2004-08-19 | Itzhak Levy | Serial advanced technology attachment interface |
WO2004075172A3 (en) * | 2003-02-14 | 2005-03-10 | Arco Comp Products Llc | Serial advanced technology attachment interface |
US20060190689A1 (en) * | 2003-03-25 | 2006-08-24 | Koninklijke Philips Electronics N.V. | Method of addressing data in a shared memory by means of an offset |
WO2004104846A2 (en) * | 2003-05-26 | 2004-12-02 | Koninklijke Philips Electronics N.V. | Method and device for transferring data between a main memory and a storage device |
WO2004104846A3 (en) * | 2003-05-26 | 2005-01-13 | Koninkl Philips Electronics Nv | Method and device for transferring data between a main memory and a storage device |
US20050005044A1 (en) * | 2003-07-02 | 2005-01-06 | Ling-Yi Liu | Storage virtualization computer system and external controller therefor |
US10452270B2 (en) | 2003-07-02 | 2019-10-22 | Infortrend Technology, Inc. | Storage virtualization computer system and external controller therefor |
CN1320436C (en) * | 2003-07-02 | 2007-06-06 | 普安科技股份有限公司 | Storage virtualization computer system and external controller therefor |
US9594510B2 (en) * | 2003-07-02 | 2017-03-14 | Infortrend Technology, Inc. | JBOD subsystem and external emulation controller thereof |
US8301809B2 (en) * | 2003-07-02 | 2012-10-30 | Infortrend Technology, Inc. | Storage virtualization computer system and external controller thereof |
US20050005062A1 (en) * | 2003-07-02 | 2005-01-06 | Ling-Yi Liu | Redundant external storage virtualization computer system |
US20050005063A1 (en) * | 2003-07-02 | 2005-01-06 | Ling-Yi Liu | Jbod subsystem and external emulation controller thereof |
CN100334567C (en) * | 2003-07-02 | 2007-08-29 | 普安科技股份有限公司 | Redundant external storage virtualization computer system |
CN1320437C (en) * | 2003-07-02 | 2007-06-06 | 普安科技股份有限公司 | Jbod subsystem and external emulation controller thereof |
US7281072B2 (en) * | 2003-07-02 | 2007-10-09 | Infortrend Technology, Inc. | Redundant external storage virtualization computer system |
US20070067537A1 (en) * | 2003-12-18 | 2007-03-22 | Pak-Lung Seto | Multiple interfaces in a storage enclosure |
US7376147B2 (en) | 2003-12-18 | 2008-05-20 | Intel Corporation | Adaptor supporting different protocols |
US20050138154A1 (en) * | 2003-12-18 | 2005-06-23 | Intel Corporation | Enclosure management device |
US20050138191A1 (en) * | 2003-12-18 | 2005-06-23 | Pak-Lung Seto | Adaptor supporting different protocols |
US20070233916A1 (en) * | 2003-12-18 | 2007-10-04 | Pak-Lung Seto | Multiple interfaces in a storage enclosure |
US7373443B2 (en) | 2003-12-18 | 2008-05-13 | Intel Corporation | Multiple interfaces in a storage enclosure |
US7373442B2 (en) * | 2003-12-18 | 2008-05-13 | Intel Corporation | Method for using an expander to connect to different storage interconnect architectures |
US7650468B2 (en) | 2003-12-22 | 2010-01-19 | Kabushiki Kaisha Kawai Gakki Seisakusho | Device for processing access concurrence to shared memory |
US20070113024A1 (en) * | 2003-12-22 | 2007-05-17 | Kabushiki Kaisha Kawai Gakki | Device for processing access concurrence to shared memory |
US7415653B1 (en) | 2004-04-21 | 2008-08-19 | Sun Microsystems, Inc. | Method and apparatus for vectored block-level checksum for file system data integrity |
US7424574B1 (en) * | 2004-04-21 | 2008-09-09 | Sun Microsystems, Inc. | Method and apparatus for dynamic striping |
US7603568B1 (en) | 2004-04-21 | 2009-10-13 | Sun Microsystems, Inc. | Method and apparatus for self-validating checksums in a file system |
US7188212B2 (en) * | 2004-05-06 | 2007-03-06 | International Business Machines Corporation | Method and system for storing data in an array of storage devices with additional and autonomic protection |
US7412450B1 (en) | 2004-05-26 | 2008-08-12 | Sun Microsystems, Inc. | Method and apparatus for identifying tampering of data in a file system |
US7281188B1 (en) | 2004-05-26 | 2007-10-09 | Sun Microsystems, Inc. | Method and system for detecting and correcting data errors using data permutations |
US7496586B1 (en) | 2004-05-26 | 2009-02-24 | Sun Microsystems, Inc. | Method and apparatus for compressing data in a file system |
US7526622B1 (en) | 2004-05-26 | 2009-04-28 | Sun Microsystems, Inc. | Method and system for detecting and correcting data errors using checksums and replication |
US7437528B1 (en) | 2004-08-17 | 2008-10-14 | Sun Microsystems, Inc. | Gang blocks |
US7533225B1 (en) | 2004-08-17 | 2009-05-12 | Sun Microsystems, Inc. | Method and apparatus for enabling adaptive endianness |
US8301810B2 (en) * | 2004-12-21 | 2012-10-30 | Infortrend Technology, Inc. | SAS storage virtualization controller, subsystem and system using the same, and method therefor |
US20060136666A1 (en) * | 2004-12-21 | 2006-06-22 | Ching-Te Pang | SAS storage virtualization controller, subsystem and system using the same, and method therefor |
US7418621B2 (en) * | 2005-02-24 | 2008-08-26 | Dot Hill Systems Corp. | Redundant storage array method and apparatus |
US20060190763A1 (en) * | 2005-02-24 | 2006-08-24 | Dot Hill Systems Corp. | Redundant storage array method and apparatus |
WO2007012919A2 (en) * | 2005-07-27 | 2007-02-01 | Adaptec, Inc. | Ripple queuing algorithm for a sas wide-port raid controller |
US20070028062A1 (en) * | 2005-07-27 | 2007-02-01 | Adaptec, Inc. | Ripple Queuing Algorithm for a SAS Wide-Port RAID Controller |
WO2007012919A3 (en) * | 2005-07-27 | 2007-04-05 | Adaptec Inc | Ripple queuing algorithm for a sas wide-port raid controller |
US7873799B2 (en) | 2005-11-04 | 2011-01-18 | Oracle America, Inc. | Method and system supporting per-file and per-block replication |
US8938594B2 (en) | 2005-11-04 | 2015-01-20 | Oracle America, Inc. | Method and system for metadata-based resilvering |
US7925827B2 (en) | 2005-11-04 | 2011-04-12 | Oracle America, Inc. | Method and system for dirty time logging |
US20070124659A1 (en) * | 2005-11-04 | 2007-05-31 | Sun Microsystems, Inc. | Method and system for data replication |
US7877554B2 (en) | 2005-11-04 | 2011-01-25 | Oracle America, Inc. | Method and system for block reallocation |
US20070106851A1 (en) * | 2005-11-04 | 2007-05-10 | Sun Microsystems, Inc. | Method and system supporting per-file and per-block replication |
US20070106925A1 (en) * | 2005-11-04 | 2007-05-10 | Sun Microsystems, Inc. | Method and system using checksums to repair data |
US7376758B2 (en) | 2005-11-04 | 2008-05-20 | Sun Microsystems, Inc. | I/O dependency graphs |
US20070106864A1 (en) * | 2005-11-04 | 2007-05-10 | Sun Microsystems, Inc. | Multiple replication levels with pooled devices |
US20070106868A1 (en) * | 2005-11-04 | 2007-05-10 | Sun Microsystems, Inc. | Method and system for latency-directed block allocation |
US20070106706A1 (en) * | 2005-11-04 | 2007-05-10 | Sun Microsystems, Inc. | Unlimited file system snapshots and clones |
US20070106632A1 (en) * | 2005-11-04 | 2007-05-10 | Sun Microsystems, Inc. | Method and system for object allocation using fill counts |
US20070106867A1 (en) * | 2005-11-04 | 2007-05-10 | Sun Microsystems, Inc. | Method and system for dirty time log directed resilvering |
US20070106865A1 (en) * | 2005-11-04 | 2007-05-10 | Sun Microsystems, Inc. | Method and system for using a block allocation policy |
US7480684B2 (en) | 2005-11-04 | 2009-01-20 | Sun Microsystems, Inc. | Method and system for object allocation using fill counts |
US20070106870A1 (en) * | 2005-11-04 | 2007-05-10 | Sun Microsystems, Inc. | Method and system for block reallocation |
US20070174494A1 (en) * | 2005-11-04 | 2007-07-26 | Sun Microsystems, Inc. | I/O dependency graphs |
US20070106863A1 (en) * | 2005-11-04 | 2007-05-10 | Sun Microsystems, Inc. | Method and system for storing a sparse file using fill counts |
US20070106866A1 (en) * | 2005-11-04 | 2007-05-10 | Sun Microsystems, Inc. | Method and system for metadata-based resilvering |
US7865673B2 (en) | 2005-11-04 | 2011-01-04 | Oracle America, Inc. | Multiple replication levels with pooled devices |
US8635190B2 (en) | 2005-11-04 | 2014-01-21 | Oracle America, Inc. | Method and system for pruned resilvering using a dirty time log |
US8549051B2 (en) | 2005-11-04 | 2013-10-01 | Oracle America, Inc. | Unlimited file system snapshots and clones |
US7596739B2 (en) | 2005-11-04 | 2009-09-29 | Sun Microsystems, Inc. | Method and system for data replication |
US20070106677A1 (en) * | 2005-11-04 | 2007-05-10 | Sun Microsystems, Inc. | Method and system for pruned resilvering using a dirty time log |
US20070106869A1 (en) * | 2005-11-04 | 2007-05-10 | Sun Microsystems, Inc. | Method and system for dirty time logging |
US8495010B2 (en) | 2005-11-04 | 2013-07-23 | Oracle America, Inc. | Method and system for adaptive metadata replication |
US20070198889A1 (en) * | 2005-11-04 | 2007-08-23 | Sun Microsystems, Inc. | Method and system for repairing partially damaged blocks |
US20070118576A1 (en) * | 2005-11-04 | 2007-05-24 | Sun Microsystems, Inc. | Method and system for adaptive metadata replication |
US20070112895A1 (en) * | 2005-11-04 | 2007-05-17 | Sun Microsystems, Inc. | Block-based incremental backup |
US7689877B2 (en) | 2005-11-04 | 2010-03-30 | Sun Microsystems, Inc. | Method and system using checksums to repair data |
US7716519B2 (en) | 2005-11-04 | 2010-05-11 | Oracle America, Inc. | Method and system for repairing partially damaged blocks |
US7899989B2 (en) | 2005-11-04 | 2011-03-01 | Oracle America, Inc. | Method and system for using a block allocation policy |
US7743225B2 (en) | 2005-11-04 | 2010-06-22 | Oracle America, Inc. | Ditto blocks |
US7930495B2 (en) | 2005-11-04 | 2011-04-19 | Oracle America, Inc. | Method and system for dirty time log directed resilvering |
US7716445B2 (en) | 2005-11-04 | 2010-05-11 | Oracle America, Inc. | Method and system for storing a sparse file using fill counts |
US20070106862A1 (en) * | 2005-11-04 | 2007-05-10 | Sun Microsystems, Inc. | Ditto blocks |
US7835359B2 (en) | 2005-12-08 | 2010-11-16 | International Business Machines Corporation | Method and apparatus for striping message payload data over a network |
US20070133536A1 (en) * | 2005-12-08 | 2007-06-14 | International Business Machines Corporation | Method and apparatus for striping message payload data over a network |
US20080104204A1 (en) * | 2006-10-31 | 2008-05-01 | Sun Microsystems, Inc. | Method and apparatus for power-managing storage devices in a storage pool |
US20080104150A1 (en) * | 2006-10-31 | 2008-05-01 | Sun Microsystems, Inc. | Method and system for priority-based allocation in a storage pool |
US7840657B2 (en) | 2006-10-31 | 2010-11-23 | Oracle America, Inc. | Method and apparatus for power-managing storage devices in a storage pool |
US7584229B2 (en) | 2006-10-31 | 2009-09-01 | Sun Microsystems, Inc. | Method and system for priority-based allocation in a storage pool |
US20080104355A1 (en) * | 2006-10-31 | 2008-05-01 | Sun Microsystems, Inc. | Method and system for reallocating blocks in a storage pool |
US7783847B2 (en) | 2006-10-31 | 2010-08-24 | Oracle America Inc. | Method and system for reallocating blocks in a storage pool |
US8645623B1 (en) * | 2007-06-28 | 2014-02-04 | Emc Corporation | Method for performing a raid operation in a data storage system |
US20090089343A1 (en) * | 2007-09-27 | 2009-04-02 | Sun Microsystems, Inc. | Method and system for block allocation for hybrid drives |
US7792882B2 (en) | 2007-09-27 | 2010-09-07 | Oracle America, Inc. | Method and system for block allocation for hybrid drives |
US20090198885A1 (en) * | 2008-02-04 | 2009-08-06 | Manoj Jose K | System and methods for host software stripe management in a striped storage subsystem |
US20090235278A1 (en) * | 2008-03-14 | 2009-09-17 | Prakash Babu H | Method for tracking and/or verifying message passing in a simulation environment |
US8387067B2 (en) * | 2008-03-14 | 2013-02-26 | Lsi Corporation | Method for tracking and/or verifying message passing in a simulation environment |
US8095728B2 (en) | 2008-04-18 | 2012-01-10 | Oracle America, Inc. | Method and system for power aware I/O scheduling |
US20090265519A1 (en) * | 2008-04-18 | 2009-10-22 | Sun Microsystems, Inc. | Method and system for power aware i/o scheduling |
US8037279B2 (en) | 2008-06-12 | 2011-10-11 | Oracle America, Inc. | Method and system for cross-domain data sharing |
US20090313446A1 (en) * | 2008-06-12 | 2009-12-17 | Sun Microsystems, Inc. | Method and system for cross-domain data sharing |
US20090327602A1 (en) * | 2008-06-30 | 2009-12-31 | Sun Microsystems, Inc. | Method and system for managing wear-level aware file systems |
US8135907B2 (en) | 2008-06-30 | 2012-03-13 | Oracle America, Inc. | Method and system for managing wear-level aware file systems |
US8280858B2 (en) | 2009-06-29 | 2012-10-02 | Oracle America, Inc. | Storage pool scrubbing with concurrent snapshots |
US20100332446A1 (en) * | 2009-06-29 | 2010-12-30 | Sun Microsystems, Inc. | Storage pool scrubbing with concurrent snapshots |
US11083381B2 (en) | 2009-09-11 | 2021-08-10 | University Of Virginia Patent Foundation | Systems and methods for determining pressure frequency changes in a subject |
US20140244938A1 (en) * | 2013-02-27 | 2014-08-28 | Vmware, Inc. | Method and Apparatus for Returning Reads in the Presence of Partial Data Unavailability |
US9298375B2 (en) * | 2013-02-27 | 2016-03-29 | Vmware, Inc. | Method and apparatus for returning reads in the presence of partial data unavailability |
US9229894B2 (en) * | 2013-04-09 | 2016-01-05 | Apple Inc. | Protocol conversion involving multiple virtual channels |
US20140304441A1 (en) * | 2013-04-09 | 2014-10-09 | Apple Inc. | Protocol conversion involving multiple virtual channels |
CN103617136A (en) * | 2013-12-04 | 2014-03-05 | 华为技术有限公司 | SCSI drive side and I/O request control method |
US10522199B2 (en) * | 2015-02-06 | 2019-12-31 | Micron Technology, Inc. | Apparatuses and methods for scatter and gather |
US11482260B2 (en) * | 2015-02-06 | 2022-10-25 | Micron Technology, Inc. | Apparatuses and methods for scatter and gather |
US10964358B2 (en) * | 2015-02-06 | 2021-03-30 | Micron Technology, Inc. | Apparatuses and methods for scatter and gather |
US9778858B1 (en) * | 2015-02-11 | 2017-10-03 | Microsemi Solutions (U.S.), Inc. | Apparatus and method for scatter gather list handling for an out of order system |
US20170048320A1 (en) * | 2015-08-13 | 2017-02-16 | Advanced Micro Devices, Inc. | Distributed gather/scatter operations across a network of memory nodes |
US10805392B2 (en) * | 2015-08-13 | 2020-10-13 | Advanced Micro Devices, Inc. | Distributed gather/scatter operations across a network of memory nodes |
US10579282B1 (en) * | 2016-03-30 | 2020-03-03 | EMC IP Holding Company LLC | Distributed copy in multi-copy replication where offset and size of I/O requests to replication site is half offset and size of I/O request to production volume |
US11960777B2 (en) | 2017-06-12 | 2024-04-16 | Pure Storage, Inc. | Utilizing multiple redundancy schemes within a unified storage element |
US11609718B1 (en) | 2017-06-12 | 2023-03-21 | Pure Storage, Inc. | Identifying valid data after a storage system recovery |
US11593036B2 (en) | 2017-06-12 | 2023-02-28 | Pure Storage, Inc. | Staging data within a unified storage element |
US11592991B2 (en) * | 2017-09-07 | 2023-02-28 | Pure Storage, Inc. | Converting raid data between persistent storage types |
US20190073141A1 (en) * | 2017-09-07 | 2019-03-07 | Pure Storage, Inc. | Converting raid data between persistent storage types |
US11775430B1 (en) * | 2018-03-12 | 2023-10-03 | Amazon Technologies, Inc. | Memory access for multiple circuit components |
CN108763023A (en) * | 2018-05-29 | 2018-11-06 | 郑州云海信息技术有限公司 | A kind of stage division of disk, device, equipment and readable storage medium storing program for executing |
US11093180B2 (en) * | 2019-09-27 | 2021-08-17 | Dell Products L.P. | RAID storage multi-operation command system |
CN112988623A (en) * | 2019-12-17 | 2021-06-18 | 北京忆芯科技有限公司 | Method and storage device for accelerating SGL (secure gateway) processing |
US11314637B2 (en) * | 2020-05-29 | 2022-04-26 | Hewlett Packard Enterprise Development Lp | System and method for efficient cache coherency protocol processing |
CN113742258A (en) * | 2020-05-29 | 2021-12-03 | 慧与发展有限责任合伙企业 | System and method for efficient cache coherency protocol processing |
Also Published As
Publication number | Publication date |
---|---|
US7155569B2 (en) | 2006-12-26 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US7155569B2 (en) | Method for raid striped I/O request generation using a shared scatter gather list | |
US8037244B2 (en) | Storage apparatus and data storage method using the same | |
US9720596B1 (en) | Coalescing writes for improved storage utilization | |
US5315602A (en) | Optimized stripe detection for redundant arrays of disk drives | |
US8156282B1 (en) | System and method for optimizing write operations in storage systems | |
US8694703B2 (en) | Hardware-accelerated lossless data compression | |
US8560879B1 (en) | Data recovery for failed memory device of memory device array | |
US9401967B2 (en) | Inline wire speed deduplication system | |
US8209587B1 (en) | System and method for eliminating zeroing of disk drives in RAID arrays | |
US6041386A (en) | Data sharing between system using different data storage formats | |
US6647460B2 (en) | Storage device with I/O counter for partial data reallocation | |
US20020087786A1 (en) | Method, system, and data structures for superimposing data records in a first data format to memory in a second data format | |
US20090034377A1 (en) | System and method for efficient updates of sequential block storage | |
US20020069318A1 (en) | Real time application accelerator and method of operating the same | |
US20090006904A1 (en) | Apparatus and method to check data integrity when handling data | |
JP2006260582A (en) | File allocating method of file system integrated with raid disk sub-system | |
KR20030034577A (en) | Stripping system, mapping and processing method thereof | |
US11256447B1 (en) | Multi-BCRC raid protection for CKD | |
US7133965B2 (en) | Raid storage device | |
Chan et al. | Elastic parity logging for SSD RAID arrays: Design, analysis, and implementation | |
US11526447B1 (en) | Destaging multiple cache slots in a single back-end track in a RAID subsystem | |
JP5141234B2 (en) | Hierarchical storage control device, hierarchical storage control system, hierarchical storage control method used therefor, and program thereof | |
JPH09288547A (en) | Array-type storage device | |
US10929032B1 (en) | Host hinting for smart disk allocation to improve sequential access performance | |
EP1110219A1 (en) | Improved method for providing variable sector-format operation to a computer system |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: LSI LOGIC CORPORATION, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:JOHNSON, STEPHEN B.;BESMER, BRAD D.;REEL/FRAME:013065/0671 Effective date: 20020627 |
|
FEPP | Fee payment procedure |
Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
FPAY | Fee payment |
Year of fee payment: 4 |
|
FPAY | Fee payment |
Year of fee payment: 8 |
|
AS | Assignment |
Owner name: LSI CORPORATION, CALIFORNIA Free format text: CHANGE OF NAME;ASSIGNOR:LSI LOGIC CORPORATION;REEL/FRAME:033102/0270 Effective date: 20070406 |
|
AS | Assignment |
Owner name: AVAGO TECHNOLOGIES GENERAL IP (SINGAPORE) PTE. LTD Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:LSI CORPORATION;REEL/FRAME:035390/0388 Effective date: 20140814 |
|
AS | Assignment |
Owner name: BANK OF AMERICA, N.A., AS COLLATERAL AGENT, NORTH CAROLINA Free format text: PATENT SECURITY AGREEMENT;ASSIGNOR:AVAGO TECHNOLOGIES GENERAL IP (SINGAPORE) PTE. LTD.;REEL/FRAME:037808/0001 Effective date: 20160201 Owner name: BANK OF AMERICA, N.A., AS COLLATERAL AGENT, NORTH Free format text: PATENT SECURITY AGREEMENT;ASSIGNOR:AVAGO TECHNOLOGIES GENERAL IP (SINGAPORE) PTE. LTD.;REEL/FRAME:037808/0001 Effective date: 20160201 |
|
AS | Assignment |
Owner name: AVAGO TECHNOLOGIES GENERAL IP (SINGAPORE) PTE. LTD., SINGAPORE Free format text: TERMINATION AND RELEASE OF SECURITY INTEREST IN PATENTS;ASSIGNOR:BANK OF AMERICA, N.A., AS COLLATERAL AGENT;REEL/FRAME:041710/0001 Effective date: 20170119 Owner name: AVAGO TECHNOLOGIES GENERAL IP (SINGAPORE) PTE. LTD Free format text: TERMINATION AND RELEASE OF SECURITY INTEREST IN PATENTS;ASSIGNOR:BANK OF AMERICA, N.A., AS COLLATERAL AGENT;REEL/FRAME:041710/0001 Effective date: 20170119 |
|
FEPP | Fee payment procedure |
Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.) |
|
AS | Assignment |
Owner name: AVAGO TECHNOLOGIES INTERNATIONAL SALES PTE. LIMITE Free format text: MERGER;ASSIGNOR:AVAGO TECHNOLOGIES GENERAL IP (SINGAPORE) PTE. LTD.;REEL/FRAME:047642/0417 Effective date: 20180509 |
|
LAPS | Lapse for failure to pay maintenance fees |
Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
STCH | Information on status: patent discontinuation |
Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362 |
|
FP | Lapsed due to failure to pay maintenance fee |
Effective date: 20181226 |
|
AS | Assignment |
Owner name: AVAGO TECHNOLOGIES INTERNATIONAL SALES PTE. LIMITE Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE EXECUTION DATE OF THE MERGER PREVIOUSLY RECORDED ON REEL 047642 FRAME 0417. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT,;ASSIGNOR:AVAGO TECHNOLOGIES GENERAL IP (SINGAPORE) PTE. LTD.;REEL/FRAME:048521/0395 Effective date: 20180905 |