Lines Matching refs:size

91     assert self.clobbered_blocks.size() == 0
208 return (sum(sr.size() for (_, sr) in self.stash_before) -
209 sum(sr.size() for (_, sr) in self.use_stash))
242 # blocksize: the size in bytes of a block, currently must be 4096.
244 # total_blocks: the total size of the partition/image, in blocks.
337 # Ensure the runtime stash size is under the limit.
349 """Limit the size of operand in command 'new' and 'zero' to 1024 blocks.
351 This prevents the target size of one command from being too large; and
360 total += blocks_to_write.size()
382 stashed_blocks += sr.size()
398 size = xf.src_ranges.size()
399 src_str = [str(size)]
413 free_size += sr.size()
422 self.AssertPartition(RangeSet(data=(0, size)), mapped_stashes)
425 self.AssertPartition(RangeSet(data=(0, size)), mapped_stashes)
437 tgt_size = xf.tgt_ranges.size()
445 assert xf.src_ranges.size() == tgt_size
449 temp_stash_usage = stashed_blocks + xf.src_ranges.size()
466 temp_stash_usage = stashed_blocks + xf.src_ranges.size()
482 assert WriteSplitTransfers(out, xf.style, to_zero) == to_zero.size()
483 total += to_zero.size()
493 # the allowed size (cache_size * threshold). There are two purposes
501 'Stash size %d (%d * %d) exceeds the limit %d (%d * %.2f)' % (
511 self.tgt.extended.size())
512 total += self.tgt.extended.size()
555 print("Revising stash size...")
569 # Compute the maximum blocks available for stash based on /cache size and
580 # Now go through all the commands. Compute the required stash size on the
593 stashed_blocks_after += sr.size()
600 print("%10d %9s %s" % (sr.size(), "explicit", use_cmd))
614 if stashed_blocks + xf.src_ranges.size() > max_allowed:
616 print("%10d %9s %s" % (xf.src_ranges.size(), "implicit", xf))
629 new_blocks += cmd.tgt_ranges.size()
638 stashed_blocks -= sr.size()
643 "insufficient cache size." % (new_blocks, num_of_bytes))
653 tgt_size = xf.tgt_ranges.size() * self.tgt.blocksize
660 tgt_size = xf.tgt_ranges.size() * self.tgt.blocksize
678 tgt_size = xf.tgt_ranges.size() * self.tgt.blocksize
800 tgt_size = xf.tgt_ranges.size() * self.tgt.blocksize
893 size = xf.src_ranges.size()
911 lost = size - xf.src_ranges.size()
957 stash_size += overlap.size()
1118 size = 0
1120 size = i.size()
1121 b.goes_before[a] = size
1122 a.goes_after[b] = size
1132 However, with the growth of file size and the shrink of the cache
1137 The downside is that after splitting, we may increase the package size
1139 1/8 of the cache size as the per-piece limit appears to be optimal.
1141 size by 30% for volantis, and 20% for angler and bullhead."""
1151 if (tgt_ranges.size() <= max_blocks_per_transfer and
1152 src_ranges.size() <= max_blocks_per_transfer):
1158 while (tgt_ranges.size() > max_blocks_per_transfer and
1159 src_ranges.size() > max_blocks_per_transfer):
1174 if tgt_ranges.size() or src_ranges.size():
1176 assert tgt_ranges.size() and src_ranges.size()
1198 # blocks while patching. We limit the analysis to files without size
1202 tgt_ranges.size() == src_ranges.size()):
1211 tgt_size = tgt_ranges.size()
1221 tgt_changed += tgt_rs.size()
1228 assert tgt_changed + tgt_skipped.size() == tgt_size
1229 print('%10d %10d (%6.2f%%) %s' % (tgt_skipped.size(), tgt_size,
1230 tgt_skipped.size() * 100.0 / tgt_size, tgt_name))