/* * scsi_merge.c Copyright (C) 1999 Eric Youngdale * * SCSI queueing library. * Initial versions: Eric Youngdale (eric@andante.org). * Based upon conversations with large numbers * of people at Linux Expo. * Support for dynamic DMA mapping: Jakub Jelinek (jakub@redhat.com). */ /* * This file contains queue management functions that are used by SCSI. * Typically this is used for several purposes. First, we need to ensure * that commands do not grow so large that they cannot be handled all at * once by a host adapter. The various flavors of merge functions included * here serve this purpose. * * Note that it would be quite trivial to allow the low-level driver the * flexibility to define it's own queue handling functions. For the time * being, the hooks are not present. Right now we are just using the * data in the host template as an indicator of how we should be handling * queues, and we select routines that are optimized for that purpose. * * Some hosts do not impose any restrictions on the size of a request. * In such cases none of the merge functions in this file are called, * and we allow ll_rw_blk to merge requests in the default manner. * This isn't guaranteed to be optimal, but it should be pretty darned * good. If someone comes up with ideas of better ways of managing queues * to improve on the default behavior, then certainly fit it into this * scheme in whatever manner makes the most sense. Please note that * since each device has it's own queue, we have considerable flexibility * in queue management. */ #define __NO_VERSION__ #include #include #include #include #include #include #include #include #include #include #include #include #include #define __KERNEL_SYSCALLS__ #include #include #include #include #include #include "scsi.h" #include "hosts.h" #include "constants.h" #include /* * This means that bounce buffers cannot be allocated in chunks > PAGE_SIZE. * Ultimately we should get away from using a dedicated DMA bounce buffer * pool, and we should instead try and use kmalloc() instead. If we can * eliminate this pool, then this restriction would no longer be needed. */ #define DMA_SEGMENT_SIZE_LIMITED #ifdef CONFIG_SCSI_DEBUG_QUEUES /* * Enable a bunch of additional consistency checking. Turn this off * if you are benchmarking. */ static int dump_stats(struct request *req, int use_clustering, int dma_host, int segments) { struct buffer_head *bh; /* * Dump the information that we have. We know we have an * inconsistency. */ printk("nr_segments is %x\n", req->nr_segments); printk("counted segments is %x\n", segments); printk("Flags %d %d\n", use_clustering, dma_host); for (bh = req->bh; bh->b_reqnext != NULL; bh = bh->b_reqnext) { printk("Segment 0x%p, blocks %d, addr 0x%lx\n", bh, bh->b_size >> 9, virt_to_phys(bh->b_data - 1)); } panic("Ththththaats all folks. Too dangerous to continue.\n"); } /* * Simple sanity check that we will use for the first go around * in order to ensure that we are doing the counting correctly. * This can be removed for optimization. */ #define SANITY_CHECK(req, _CLUSTER, _DMA) \ if( req->nr_segments != __count_segments(req, _CLUSTER, _DMA, NULL) ) \ { \ printk("Incorrect segment count at 0x%p", current_text_addr()); \ dump_stats(req, _CLUSTER, _DMA, __count_segments(req, _CLUSTER, _DMA, NULL)); \ } #else #define SANITY_CHECK(req, _CLUSTER, _DMA) #endif static void dma_exhausted(Scsi_Cmnd * SCpnt, int i) { int jj; struct scatterlist *sgpnt; void **bbpnt; int consumed = 0; sgpnt = (struct scatterlist *) SCpnt->request_buffer; bbpnt = SCpnt->bounce_buffers; /* * Now print out a bunch of stats. First, start with the request * size. */ printk("dma_free_sectors:%d\n", scsi_dma_free_sectors); printk("use_sg:%d\ti:%d\n", SCpnt->use_sg, i); printk("request_bufflen:%d\n", SCpnt->request_bufflen); /* * Now dump the scatter-gather table, up to the point of failure. */ for(jj=0; jj < SCpnt->use_sg; jj++) { printk("[%d]\tlen:%d\taddr:%p\tbounce:%p\n", jj, sgpnt[jj].length, sgpnt[jj].address, (bbpnt ? bbpnt[jj] : NULL)); if (bbpnt && bbpnt[jj]) consumed += sgpnt[jj].length; } printk("Total %d sectors consumed\n", consumed); panic("DMA pool exhausted"); } #define CLUSTERABLE_DEVICE(SH,SD) (SH->use_clustering) /* * This entire source file deals with the new queueing code. */ /* * Function: __count_segments() * * Purpose: Prototype for queue merge function. * * Arguments: q - Queue for which we are merging request. * req - request into which we wish to merge. * use_clustering - 1 if this host wishes to use clustering * dma_host - 1 if this host has ISA DMA issues (bus doesn't * expose all of the address lines, so that DMA cannot * be done from an arbitrary address). * remainder - used to track the residual size of the last * segment. Comes in handy when we want to limit the * size of bounce buffer segments to PAGE_SIZE. * * Returns: Count of the number of SG segments for the request. * * Lock status: * * Notes: This is only used for diagnostic purposes. */ __inline static int __count_segments(struct request *req, int use_clustering, int dma_host, int * remainder) { int ret = 1; int reqsize = 0; struct buffer_head *bh; struct buffer_head *bhnext; if( remainder != NULL ) { reqsize = *remainder; } /* * Add in the size increment for the first buffer. */ bh = req->bh; #ifdef DMA_SEGMENT_SIZE_LIMITED if( reqsize + bh->b_size > PAGE_SIZE ) { ret++; reqsize = bh->b_size; } else { reqsize += bh->b_size; } #else reqsize += bh->b_size; #endif for (bh = req->bh, bhnext = bh->b_reqnext; bhnext != NULL; bh = bhnext, bhnext = bh->b_reqnext) { if (use_clustering) { /* * See if we can do this without creating another * scatter-gather segment. In the event that this is a * DMA capable host, make sure that a segment doesn't span * the DMA threshold boundary. */ if (dma_host && virt_to_phys(bhnext->b_data) - 1 == ISA_DMA_THRESHOLD) { ret++; reqsize = bhnext->b_size; } else if (CONTIGUOUS_BUFFERS(bh, bhnext)) { /* * This one is OK. Let it go. */ #ifdef DMA_SEGMENT_SIZE_LIMITED /* Note scsi_malloc is only able to hand out * chunks of memory in sizes of PAGE_SIZE or * less. Thus we need to keep track of * the size of the piece that we have * seen so far, and if we have hit * the limit of PAGE_SIZE, then we are * kind of screwed and we need to start * another segment. */ if( dma_host && virt_to_phys(bh->b_data) - 1 >= ISA_DMA_THRESHOLD && reqsize + bhnext->b_size > PAGE_SIZE ) { ret++; reqsize = bhnext->b_size; continue; } #endif reqsize += bhnext->b_size; continue; } ret++; reqsize = bhnext->b_size; } else { ret++; reqsize = bhnext->b_size; } } if( remainder != NULL ) { *remainder = reqsize; } return ret; } /* * Function: recount_segments() * * Purpose: Recount the number of scatter-gather segments for this request. * * Arguments: req - request that needs recounting. * * Returns: Count of the number of SG segments for the request. * * Lock status: Irrelevant. * * Notes: This is only used when we have partially completed requests * and the bit that is leftover is of an indeterminate size. * This can come up if you get a MEDIUM_ERROR, for example, * as we will have "completed" all of the sectors up to and * including the bad sector, and the leftover bit is what * we have to do now. This tends to be a rare occurrence, so * we aren't busting our butts to instantiate separate versions * of this function for the 4 different flag values. We * probably should, however. */ void recount_segments(Scsi_Cmnd * SCpnt) { struct request *req; struct Scsi_Host *SHpnt; Scsi_Device * SDpnt; req = &SCpnt->request; SHpnt = SCpnt->host; SDpnt = SCpnt->device; req->nr_segments = __count_segments(req, CLUSTERABLE_DEVICE(SHpnt, SDpnt), SHpnt->unchecked_isa_dma, NULL); } #define MERGEABLE_BUFFERS(X,Y) \ (((((long)(X)->b_data+(X)->b_size)|((long)(Y)->b_data)) & \ (DMA_CHUNK_SIZE - 1)) == 0) #ifdef DMA_CHUNK_SIZE static inline int scsi_new_mergeable(request_queue_t * q, struct request * req, struct Scsi_Host *SHpnt, int max_segments) { /* * pci_map_sg will be able to merge these two * into a single hardware sg entry, check if * we'll have enough memory for the sg list. * scsi.c allocates for this purpose * min(64,sg_tablesize) entries. */ if (req->nr_segments >= max_segments || req->nr_segments >= SHpnt->sg_tablesize) return 0; req->nr_segments++; return 1; } static inline int scsi_new_segment(request_queue_t * q, struct request * req, struct Scsi_Host *SHpnt, int max_segments) { /* * pci_map_sg won't be able to map these two * into a single hardware sg entry, so we have to * check if things fit into sg_tablesize. */ if (req->nr_hw_segments >= SHpnt->sg_tablesize || req->nr_segments >= SHpnt->sg_tablesize) return 0; req->nr_hw_segments++; req->nr_segments++; return 1; } #else static inline int scsi_new_segment(request_queue_t * q, struct request * req, struct Scsi_Host *SHpnt, int max_segments) { if (req->nr_segments < SHpnt->sg_tablesize && req->nr_segments < max_segments) { /* * This will form the start of a new segment. Bump the * counter. */ req->nr_segments++; return 1; } else { return 0; } } #endif /* * Function: __scsi_merge_fn() * * Purpose: Prototype for queue merge function. * * Arguments: q - Queue for which we are merging request. * req - request into which we wish to merge. * bh - Block which we may wish to merge into request * use_clustering - 1 if this host wishes to use clustering * dma_host - 1 if this host has ISA DMA issues (bus doesn't * expose all of the address lines, so that DMA cannot * be done from an arbitrary address). * * Returns: 1 if it is OK to merge the block into the request. 0 * if it is not OK. * * Lock status: io_request_lock is assumed to be held here. * * Notes: Some drivers have limited scatter-gather table sizes, and * thus they cannot queue an infinitely large command. This * function is called from ll_rw_blk before it attempts to merge * a new block into a request to make sure that the request will * not become too large. * * This function is not designed to be directly called. Instead * it should be referenced from other functions where the * use_clustering and dma_host parameters should be integer * constants. The compiler should thus be able to properly * optimize the code, eliminating stuff that is irrelevant. * It is more maintainable to do this way with a single function * than to have 4 separate functions all doing roughly the * same thing. */ __inline static int __scsi_back_merge_fn(request_queue_t * q, struct request *req, struct buffer_head *bh, int max_segments, int use_clustering, int dma_host) { unsigned int count; unsigned int segment_size = 0; Scsi_Device *SDpnt; struct Scsi_Host *SHpnt; SDpnt = (Scsi_Device *) q->queuedata; SHpnt = SDpnt->host; #ifdef DMA_CHUNK_SIZE if (max_segments > 64) max_segments = 64; #endif if ((req->nr_sectors + (bh->b_size >> 9)) > SHpnt->max_sectors) return 0; if (use_clustering) { /* * See if we can do this without creating another * scatter-gather segment. In the event that this is a * DMA capable host, make sure that a segment doesn't span * the DMA threshold boundary. */ if (dma_host && virt_to_phys(req->bhtail->b_data) - 1 == ISA_DMA_THRESHOLD) { goto new_end_segment; } if (CONTIGUOUS_BUFFERS(req->bhtail, bh)) { #ifdef DMA_SEGMENT_SIZE_LIMITED if( dma_host && virt_to_phys(bh->b_data) - 1 >= ISA_DMA_THRESHOLD ) { segment_size = 0; count = __count_segments(req, use_clustering, dma_host, &segment_size); if( segment_size + bh->b_size > PAGE_SIZE ) { goto new_end_segment; } } #endif /* * This one is OK. Let it go. */ return 1; } } new_end_segment: #ifdef DMA_CHUNK_SIZE if (MERGEABLE_BUFFERS(req->bhtail, bh)) return scsi_new_mergeable(q, req, SHpnt, max_segments); #endif return scsi_new_segment(q, req, SHpnt, max_segments); } __inline static int __scsi_front_merge_fn(request_queue_t * q, struct request *req, struct buffer_head *bh, int max_segments, int use_clustering, int dma_host) { unsigned int count; unsigned int segment_size = 0; Scsi_Device *SDpnt; struct Scsi_Host *SHpnt; SDpnt = (Scsi_Device *) q->queuedata; SHpnt = SDpnt->host; #ifdef DMA_CHUNK_SIZE if (max_segments > 64) max_segments = 64; #endif if ((req->nr_sectors + (bh->b_size >> 9)) > SHpnt->max_sectors) return 0; if (use_clustering) { /* * See if we can do this without creating another * scatter-gather segment. In the event that this is a * DMA capable host, make sure that a segment doesn't span * the DMA threshold boundary. */ if (dma_host && virt_to_phys(bh->b_data) - 1 == ISA_DMA_THRESHOLD) { goto new_start_segment; } if (CONTIGUOUS_BUFFERS(bh, req->bh)) { #ifdef DMA_SEGMENT_SIZE_LIMITED if( dma_host && virt_to_phys(bh->b_data) - 1 >= ISA_DMA_THRESHOLD ) { segment_size = bh->b_size; count = __count_segments(req, use_clustering, dma_host, &segment_size); if( count != req->nr_segments ) { goto new_start_segment; } } #endif /* * This one is OK. Let it go. */ return 1; } } new_start_segment: #ifdef DMA_CHUNK_SIZE if (MERGEABLE_BUFFERS(bh, req->bh)) return scsi_new_mergeable(q, req, SHpnt, max_segments); #endif return scsi_new_segment(q, req, SHpnt, max_segments); } /* * Function: scsi_merge_fn_() * * Purpose: queue merge function. * * Arguments: q - Queue for which we are merging request. * req - request into which we wish to merge. * bh - Block which we may wish to merge into request * * Returns: 1 if it is OK to merge the block into the request. 0 * if it is not OK. * * Lock status: io_request_lock is assumed to be held here. * * Notes: Optimized for different cases depending upon whether * ISA DMA is in use and whether clustering should be used. */ #define MERGEFCT(_FUNCTION, _BACK_FRONT, _CLUSTER, _DMA) \ static int _FUNCTION(request_queue_t * q, \ struct request * req, \ struct buffer_head * bh, \ int max_segments) \ { \ int ret; \ SANITY_CHECK(req, _CLUSTER, _DMA); \ ret = __scsi_ ## _BACK_FRONT ## _merge_fn(q, \ req, \ bh, \ max_segments, \ _CLUSTER, \ _DMA); \ return ret; \ } /* Version with use_clustering 0 and dma_host 1 is not necessary, * since the only use of dma_host above is protected by use_clustering. */ MERGEFCT(scsi_back_merge_fn_, back, 0, 0) MERGEFCT(scsi_back_merge_fn_c, back, 1, 0) MERGEFCT(scsi_back_merge_fn_dc, back, 1, 1) MERGEFCT(scsi_front_merge_fn_, front, 0, 0) MERGEFCT(scsi_front_merge_fn_c, front, 1, 0) MERGEFCT(scsi_front_merge_fn_dc, front, 1, 1) /* * Function: __scsi_merge_requests_fn() * * Purpose: Prototype for queue merge function. * * Arguments: q - Queue for which we are merging request. * req - request into which we wish to merge. * next - 2nd request that we might want to combine with req * use_clustering - 1 if this host wishes to use clustering * dma_host - 1 if this host has ISA DMA issues (bus doesn't * expose all of the address lines, so that DMA cannot * be done from an arbitrary address). * * Returns: 1 if it is OK to merge the two requests. 0 * if it is not OK. * * Lock status: io_request_lock is assumed to be held here. * * Notes: Some drivers have limited scatter-gather table sizes, and * thus they cannot queue an infinitely large command. This * function is called from ll_rw_blk before it attempts to merge * a new block into a request to make sure that the request will * not become too large. * * This function is not designed to be directly called. Instead * it should be referenced from other functions where the * use_clustering and dma_host parameters should be integer * constants. The compiler should thus be able to properly * optimize the code, eliminating stuff that is irrelevant. * It is more maintainable to do this way with a single function * than to have 4 separate functions all doing roughly the * same thing. */ __inline static int __scsi_merge_requests_fn(request_queue_t * q, struct request *req, struct request *next, int max_segments, int use_clustering, int dma_host) { Scsi_Device *SDpnt; struct Scsi_Host *SHpnt; /* * First check if the either of the requests are re-queued * requests. Can't merge them if they are. */ if (req->special || next->special) return 0; SDpnt = (Scsi_Device *) q->queuedata; SHpnt = SDpnt->host; #ifdef DMA_CHUNK_SIZE if (max_segments > 64) max_segments = 64; /* If it would not fit into prepared memory space for sg chain, * then don't allow the merge. */ if (req->nr_segments + next->nr_segments - 1 > max_segments || req->nr_segments + next->nr_segments - 1 > SHpnt->sg_tablesize) { return 0; } if (req->nr_hw_segments + next->nr_hw_segments - 1 > SHpnt->sg_tablesize) { return 0; } #else /* * If the two requests together are too large (even assuming that we * can merge the boundary requests into one segment, then don't * allow the merge. */ if (req->nr_segments + next->nr_segments - 1 > SHpnt->sg_tablesize) { return 0; } #endif if ((req->nr_sectors + next->nr_sectors) > SHpnt->max_sectors) return 0; /* * The main question is whether the two segments at the boundaries * would be considered one or two. */ if (use_clustering) { /* * See if we can do this without creating another * scatter-gather segment. In the event that this is a * DMA capable host, make sure that a segment doesn't span * the DMA threshold boundary. */ if (dma_host && virt_to_phys(req->bhtail->b_data) - 1 == ISA_DMA_THRESHOLD) { goto dont_combine; } #ifdef DMA_SEGMENT_SIZE_LIMITED /* * We currently can only allocate scatter-gather bounce * buffers in chunks of PAGE_SIZE or less. */ if (dma_host && CONTIGUOUS_BUFFERS(req->bhtail, next->bh) && virt_to_phys(req->bhtail->b_data) - 1 >= ISA_DMA_THRESHOLD ) { int segment_size = 0; int count = 0; count = __count_segments(req, use_clustering, dma_host, &segment_size); count += __count_segments(next, use_clustering, dma_host, &segment_size); if( count != req->nr_segments + next->nr_segments ) { goto dont_combine; } } #endif if (CONTIGUOUS_BUFFERS(req->bhtail, next->bh)) { /* * This one is OK. Let it go. */ req->nr_segments += next->nr_segments - 1; #ifdef DMA_CHUNK_SIZE req->nr_hw_segments += next->nr_hw_segments - 1; #endif return 1; } } dont_combine: #ifdef DMA_CHUNK_SIZE if (req->nr_segments + next->nr_segments > max_segments || req->nr_segments + next->nr_segments > SHpnt->sg_tablesize) { return 0; } /* If dynamic DMA mapping can merge last segment in req with * first segment in next, then the check for hw segments was * done above already, so we can always merge. */ if (MERGEABLE_BUFFERS (req->bhtail, next->bh)) { req->nr_hw_segments += next->nr_hw_segments - 1; } else if (req->nr_hw_segments + next->nr_hw_segments > SHpnt->sg_tablesize) { return 0; } else { req->nr_hw_segments += next->nr_hw_segments; } req->nr_segments += next->nr_segments; return 1; #else /* * We know that the two requests at the boundary should not be combined. * Make sure we can fix something that is the sum of the two. * A slightly stricter test than we had above. */ if (req->nr_segments + next->nr_segments > max_segments || req->nr_segments + next->nr_segments > SHpnt->sg_tablesize) { return 0; } else { /* * This will form the start of a new segment. Bump the * counter. */ req->nr_segments += next->nr_segments; return 1; } #endif } /* * Function: scsi_merge_requests_fn_() * * Purpose: queue merge function. * * Arguments: q - Queue for which we are merging request. * req - request into which we wish to merge. * bh - Block which we may wish to merge into request * * Returns: 1 if it is OK to merge the block into the request. 0 * if it is not OK. * * Lock status: io_request_lock is assumed to be held here. * * Notes: Optimized for different cases depending upon whether * ISA DMA is in use and whether clustering should be used. */ #define MERGEREQFCT(_FUNCTION, _CLUSTER, _DMA) \ static int _FUNCTION(request_queue_t * q, \ struct request * req, \ struct request * next, \ int max_segments) \ { \ int ret; \ SANITY_CHECK(req, _CLUSTER, _DMA); \ ret = __scsi_merge_requests_fn(q, req, next, max_segments, _CLUSTER, _DMA); \ return ret; \ } /* Version with use_clustering 0 and dma_host 1 is not necessary, * since the only use of dma_host above is protected by use_clustering. */ MERGEREQFCT(scsi_merge_requests_fn_, 0, 0) MERGEREQFCT(scsi_merge_requests_fn_c, 1, 0) MERGEREQFCT(scsi_merge_requests_fn_dc, 1, 1) /* * Function: __init_io() * * Purpose: Prototype for io initialize function. * * Arguments: SCpnt - Command descriptor we wish to initialize * sg_count_valid - 1 if the sg count in the req is valid. * use_clustering - 1 if this host wishes to use clustering * dma_host - 1 if this host has ISA DMA issues (bus doesn't * expose all of the address lines, so that DMA cannot * be done from an arbitrary address). * * Returns: 1 on success. * * Lock status: * * Notes: Only the SCpnt argument should be a non-constant variable. * This function is designed in such a way that it will be * invoked from a series of small stubs, each of which would * be optimized for specific circumstances. * * The advantage of this is that hosts that don't do DMA * get versions of the function that essentially don't have * any of the DMA code. Same goes for clustering - in the * case of hosts with no need for clustering, there is no point * in a whole bunch of overhead. * * Finally, in the event that a host has set can_queue to SG_ALL * implying that there is no limit to the length of a scatter * gather list, the sg count in the request won't be valid * (mainly because we don't need queue management functions * which keep the tally uptodate. */ __inline static int __init_io(Scsi_Cmnd * SCpnt, int sg_count_valid, int use_clustering, int dma_host) { struct buffer_head * bh; struct buffer_head * bhprev; char * buff; int count; int i; struct request * req; int sectors; struct scatterlist * sgpnt; int this_count; void ** bbpnt; /* * FIXME(eric) - don't inline this - it doesn't depend on the * integer flags. Come to think of it, I don't think this is even * needed any more. Need to play with it and see if we hit the * panic. If not, then don't bother. */ if (!SCpnt->request.bh) { /* * Case of page request (i.e. raw device), or unlinked buffer * Typically used for swapping, but this isn't how we do * swapping any more. */ panic("I believe this is dead code. If we hit this, I was wrong"); #if 0 SCpnt->request_bufflen = SCpnt->request.nr_sectors << 9; SCpnt->request_buffer = SCpnt->request.buffer; SCpnt->use_sg = 0; /* * FIXME(eric) - need to handle DMA here. */ #endif return 1; } req = &SCpnt->request; /* * First we need to know how many scatter gather segments are needed. */ if (!sg_count_valid) { count = __count_segments(req, use_clustering, dma_host, NULL); } else { count = req->nr_segments; } /* * If the dma pool is nearly empty, then queue a minimal request * with a single segment. Typically this will satisfy a single * buffer. */ if (dma_host && scsi_dma_free_sectors <= 10) { this_count = SCpnt->request.current_nr_sectors; goto single_segment; } /* * Don't bother with scatter-gather if there is only one segment. */ if (count == 1) { this_count = SCpnt->request.nr_sectors; goto single_segment; } SCpnt->use_sg = count; /* * Allocate the actual scatter-gather table itself. */ SCpnt->sglist_len = (SCpnt->use_sg * sizeof(struct scatterlist)); /* If we could potentially require ISA bounce buffers, allocate * space for this array here. */ if (dma_host) SCpnt->sglist_len += (SCpnt->use_sg * sizeof(void *)); /* scsi_malloc can only allocate in chunks of 512 bytes so * round it up. */ SCpnt->sglist_len = (SCpnt->sglist_len + 511) & ~511; sgpnt = (struct scatterlist *) scsi_malloc(SCpnt->sglist_len); /* * Now fill the scatter-gather table. */ if (!sgpnt) { /* * If we cannot allocate the scatter-gather table, then * simply write the first buffer all by itself. */ printk("Warning - running *really* short on DMA buffers\n"); this_count = SCpnt->request.current_nr_sectors; goto single_segment; } /* * Next, walk the list, and fill in the addresses and sizes of * each segment. */ memset(sgpnt, 0, SCpnt->sglist_len); SCpnt->request_buffer = (char *) sgpnt; SCpnt->request_bufflen = 0; bhprev = NULL; if (dma_host) bbpnt = (void **) ((char *)sgpnt + (SCpnt->use_sg * sizeof(struct scatterlist))); else bbpnt = NULL; SCpnt->bounce_buffers = bbpnt; for (count = 0, bh = SCpnt->request.bh; bh; bh = bh->b_reqnext) { if (use_clustering && bhprev != NULL) { if (dma_host && virt_to_phys(bhprev->b_data) - 1 == ISA_DMA_THRESHOLD) { /* Nothing - fall through */ } else if (CONTIGUOUS_BUFFERS(bhprev, bh)) { /* * This one is OK. Let it go. Note that we * do not have the ability to allocate * bounce buffer segments > PAGE_SIZE, so * for now we limit the thing. */ if( dma_host ) { #ifdef DMA_SEGMENT_SIZE_LIMITED if( virt_to_phys(bh->b_data) - 1 < ISA_DMA_THRESHOLD || sgpnt[count - 1].length + bh->b_size <= PAGE_SIZE ) { sgpnt[count - 1].length += bh->b_size; bhprev = bh; continue; } #else sgpnt[count - 1].length += bh->b_size; bhprev = bh; continue; #endif } else { sgpnt[count - 1].length += bh->b_size; SCpnt->request_bufflen += bh->b_size; bhprev = bh; continue; } } } count++; sgpnt[count - 1].address = bh->b_data; sgpnt[count - 1].page = NULL; sgpnt[count - 1].length += bh->b_size; if (!dma_host) { SCpnt->request_bufflen += bh->b_size; } bhprev = bh; } /* * Verify that the count is correct. */ if (count != SCpnt->use_sg) { printk("Incorrect number of segments after building list\n"); #ifdef CONFIG_SCSI_DEBUG_QUEUES dump_stats(req, use_clustering, dma_host, count); #endif } if (!dma_host) { return 1; } /* * Now allocate bounce buffers, if needed. */ SCpnt->request_bufflen = 0; for (i = 0; i < count; i++) { sectors = (sgpnt[i].length >> 9); SCpnt->request_bufflen += sgpnt[i].length; if (virt_to_phys(sgpnt[i].address) + sgpnt[i].length - 1 > ISA_DMA_THRESHOLD) { if( scsi_dma_free_sectors - sectors <= 10 ) { /* * If this would nearly drain the DMA * pool empty, then let's stop here. * Don't make this request any larger. * This is kind of a safety valve that * we use - we could get screwed later * on if we run out completely. */ SCpnt->request_bufflen -= sgpnt[i].length; SCpnt->use_sg = i; if (i == 0) { goto big_trouble; } break; } bbpnt[i] = sgpnt[i].address; sgpnt[i].address = (char *) scsi_malloc(sgpnt[i].length); /* * If we cannot allocate memory for this DMA bounce * buffer, then queue just what we have done so far. */ if (sgpnt[i].address == NULL) { printk("Warning - running low on DMA memory\n"); SCpnt->request_bufflen -= sgpnt[i].length; SCpnt->use_sg = i; if (i == 0) { goto big_trouble; } break; } if (SCpnt->request.cmd == WRITE) { memcpy(sgpnt[i].address, bbpnt[i], sgpnt[i].length); } } } return 1; big_trouble: /* * We come here in the event that we get one humongous * request, where we need a bounce buffer, and the buffer is * more than we can allocate in a single call to * scsi_malloc(). In addition, we only come here when it is * the 0th element of the scatter-gather table that gets us * into this trouble. As a fallback, we fall back to * non-scatter-gather, and ask for a single segment. We make * a half-hearted attempt to pick a reasonably large request * size mainly so that we don't thrash the thing with * iddy-biddy requests. */ /* * The original number of sectors in the 0th element of the * scatter-gather table. */ sectors = sgpnt[0].length >> 9; /* * Free up the original scatter-gather table. Note that since * it was the 0th element that got us here, we don't have to * go in and free up memory from the other slots. */ SCpnt->request_bufflen = 0; SCpnt->use_sg = 0; scsi_free(SCpnt->request_buffer, SCpnt->sglist_len); /* * Make an attempt to pick up as much as we reasonably can. * Just keep adding sectors until the pool starts running kind of * low. The limit of 30 is somewhat arbitrary - the point is that * it would kind of suck if we dropped down and limited ourselves to * single-block requests if we had hundreds of free sectors. */ if( scsi_dma_free_sectors > 30 ) { for (this_count = 0, bh = SCpnt->request.bh; bh; bh = bh->b_reqnext) { if( scsi_dma_free_sectors - this_count < 30 || this_count == sectors ) { break; } this_count += bh->b_size >> 9; } } else { /* * Yow! Take the absolute minimum here. */ this_count = SCpnt->request.current_nr_sectors; } /* * Now drop through into the single-segment case. */ single_segment: /* * Come here if for any reason we choose to do this as a single * segment. Possibly the entire request, or possibly a small * chunk of the entire request. */ bh = SCpnt->request.bh; buff = SCpnt->request.buffer; if (dma_host) { /* * Allocate a DMA bounce buffer. If the allocation fails, fall * back and allocate a really small one - enough to satisfy * the first buffer. */ if (virt_to_phys(SCpnt->request.bh->b_data) + (this_count << 9) - 1 > ISA_DMA_THRESHOLD) { buff = (char *) scsi_malloc(this_count << 9); if (!buff) { printk("Warning - running low on DMA memory\n"); this_count = SCpnt->request.current_nr_sectors; buff = (char *) scsi_malloc(this_count << 9); if (!buff) { dma_exhausted(SCpnt, 0); } } if (SCpnt->request.cmd == WRITE) memcpy(buff, (char *) SCpnt->request.buffer, this_count << 9); } } SCpnt->request_bufflen = this_count << 9; SCpnt->request_buffer = buff; SCpnt->use_sg = 0; return 1; } #define INITIO(_FUNCTION, _VALID, _CLUSTER, _DMA) \ static int _FUNCTION(Scsi_Cmnd * SCpnt) \ { \ return __init_io(SCpnt, _VALID, _CLUSTER, _DMA); \ } /* * ll_rw_blk.c now keeps track of the number of segments in * a request. Thus we don't have to do it any more here. * We always force "_VALID" to 1. Eventually clean this up * and get rid of the extra argument. */ INITIO(scsi_init_io_v, 1, 0, 0) INITIO(scsi_init_io_vd, 1, 0, 1) INITIO(scsi_init_io_vc, 1, 1, 0) INITIO(scsi_init_io_vdc, 1, 1, 1) /* * Function: initialize_merge_fn() * * Purpose: Initialize merge function for a host * * Arguments: SHpnt - Host descriptor. * * Returns: Nothing. * * Lock status: * * Notes: */ void initialize_merge_fn(Scsi_Device * SDpnt) { request_queue_t *q; struct Scsi_Host *SHpnt; SHpnt = SDpnt->host; q = &SDpnt->request_queue; /* * If the host has already selected a merge manager, then don't * pick a new one. */ #if 0 if (q->back_merge_fn && q->front_merge_fn) return; #endif /* * If this host has an unlimited tablesize, then don't bother with a * merge manager. The whole point of the operation is to make sure * that requests don't grow too large, and this host isn't picky. * * Note that ll_rw_blk.c is effectively maintaining a segment * count which is only valid if clustering is used, and it obviously * doesn't handle the DMA case. In the end, it * is simply easier to do it ourselves with our own functions * rather than rely upon the default behavior of ll_rw_blk. */ if (!CLUSTERABLE_DEVICE(SHpnt, SDpnt) && SHpnt->unchecked_isa_dma == 0) { q->back_merge_fn = scsi_back_merge_fn_; q->front_merge_fn = scsi_front_merge_fn_; q->merge_requests_fn = scsi_merge_requests_fn_; SDpnt->scsi_init_io_fn = scsi_init_io_v; } else if (!CLUSTERABLE_DEVICE(SHpnt, SDpnt) && SHpnt->unchecked_isa_dma != 0) { q->back_merge_fn = scsi_back_merge_fn_; q->front_merge_fn = scsi_front_merge_fn_; q->merge_requests_fn = scsi_merge_requests_fn_; SDpnt->scsi_init_io_fn = scsi_init_io_vd; } else if (CLUSTERABLE_DEVICE(SHpnt, SDpnt) && SHpnt->unchecked_isa_dma == 0) { q->back_merge_fn = scsi_back_merge_fn_c; q->front_merge_fn = scsi_front_merge_fn_c; q->merge_requests_fn = scsi_merge_requests_fn_c; SDpnt->scsi_init_io_fn = scsi_init_io_vc; } else if (CLUSTERABLE_DEVICE(SHpnt, SDpnt) && SHpnt->unchecked_isa_dma != 0) { q->back_merge_fn = scsi_back_merge_fn_dc; q->front_merge_fn = scsi_front_merge_fn_dc; q->merge_requests_fn = scsi_merge_requests_fn_dc; SDpnt->scsi_init_io_fn = scsi_init_io_vdc; } }