/external/icu/android_icu4j/src/main/java/android/icu/text/ |
H A D | BytesDictionaryMatcher.java | 46 public int matches(CharacterIterator text_, int maxLength, int[] lengths, int[] count_, int limit, int[] values) { argument 63 lengths[count] = numChars;
|
H A D | CharsDictionaryMatcher.java | 25 public int matches(CharacterIterator text_, int maxLength, int[] lengths, int[] count_, int limit, int[] values) { argument 42 lengths[count] = numChars;
|
H A D | CjkBreakEngine.java | 147 int lengths[] = new int[numChars]; 158 fDictionary.matches(text, maxSearchLength, lengths, count_, maxSearchLength, values); 167 if ((count == 0 || lengths[0] != 1) && current32(text) != DONE32 && !fHangulWordSet.contains(current32(text))) { 169 lengths[count] = 1; 175 if (newSnlp < bestSnlp[lengths[j] + i]) { 176 bestSnlp[lengths[j] + i] = newSnlp; 177 prev[lengths[j] + i] = i;
|
/external/icu/icu4j/main/classes/core/src/com/ibm/icu/text/ |
H A D | CharsDictionaryMatcher.java | 24 public int matches(CharacterIterator text_, int maxLength, int[] lengths, int[] count_, int limit, int[] values) { argument 41 lengths[count] = numChars;
|
H A D | CjkBreakEngine.java | 146 int lengths[] = new int[numChars]; 157 fDictionary.matches(text, maxSearchLength, lengths, count_, maxSearchLength, values); 166 if ((count == 0 || lengths[0] != 1) && current32(text) != DONE32 && !fHangulWordSet.contains(current32(text))) { 168 lengths[count] = 1; 174 if (newSnlp < bestSnlp[lengths[j] + i]) { 175 bestSnlp[lengths[j] + i] = newSnlp; 176 prev[lengths[j] + i] = i;
|
H A D | BytesDictionaryMatcher.java | 45 public int matches(CharacterIterator text_, int maxLength, int[] lengths, int[] count_, int limit, int[] values) { argument 62 lengths[count] = numChars;
|
/external/python/cpython3/Lib/test/ |
H A D | test_bufio.py | 11 lengths = list(range(1, 257)) + [512, 1000, 1024, 2048, 4096, 8192, 10000, variable 43 for length in lengths:
|
/external/toybox/toys/other/ |
H A D | vmstat.c | 83 "sy\0id\0wa", lengths[] = {2,2,6,6,6,6,4,4,5,5,4,4,2,2,2,2}; local 104 for (i=0; i<sizeof(lengths); i++) { 105 printf(" %*s"+!i, lengths[i], header); 136 for (i=0; i<sizeof(lengths); i++) { 149 expected += lengths[i] + !!i;
|
/external/brotli/c/enc/ |
H A D | histogram.c | 31 self->length_ = split->lengths ? split->lengths[0] : 0; 38 self->length_ = self->split_->lengths[self->idx_];
|
H A D | metablock_inc.h | 67 split->lengths, split->lengths_alloc_size, max_num_blocks); 93 split->lengths[0] = (uint32_t)self->block_size_; 125 split->lengths[self->num_blocks_] = (uint32_t)self->block_size_; 141 split->lengths[self->num_blocks_] = (uint32_t)self->block_size_; 154 split->lengths[self->num_blocks_ - 1] += (uint32_t)self->block_size_;
|
/external/icu/icu4c/source/common/ |
H A D | dictionarydata.h | 80 * @param lengths output array, filled with the lengths of the matches, in order, 83 * @param cpLengths output array, filled with the lengths of the matches, in order, 94 int32_t *lengths, int32_t *cpLengths, int32_t *values, 109 int32_t *lengths, int32_t *cpLengths, int32_t *values, 127 int32_t *lengths, int32_t *cpLengths, int32_t *values,
|
H A D | dictionarydata.cpp | 46 int32_t *lengths, int32_t *cpLengths, int32_t *values, 63 if (lengths != NULL) { 64 lengths[wordCount] = lengthMatched; 114 int32_t *lengths, int32_t *cpLengths, int32_t *values, 130 if (lengths != NULL) { 131 lengths[wordCount] = lengthMatched; 45 matches(UText *text, int32_t maxLength, int32_t limit, int32_t *lengths, int32_t *cpLengths, int32_t *values, int32_t *prefix) const argument 113 matches(UText *text, int32_t maxLength, int32_t limit, int32_t *lengths, int32_t *cpLengths, int32_t *values, int32_t *prefix) const argument
|
/external/python/cpython2/Lib/test/ |
H A D | test_bufio.py | 11 lengths = list(range(1, 257)) + [512, 1000, 1024, 2048, 4096, 8192, 10000, variable 43 for length in lengths:
|
/external/tensorflow/tensorflow/contrib/learn/python/learn/learn_io/ |
H A D | dask_io.py | 42 lengths = df.map_partitions(len).compute() 43 divisions = np.cumsum(lengths).tolist()
|
/external/glide/third_party/disklrucache/src/main/java/com/bumptech/glide/disklrucache/ |
H A D | DiskLruCache.java | 126 * and may be read. A publish line is followed by the lengths of each of 330 size += entry.lengths[t]; 433 return new Value(key, entry.sequenceNumber, entry.cleanFiles, entry.lengths); 527 long oldLength = entry.lengths[i]; 529 entry.lengths[i] = newLength; 592 size -= entry.lengths[i]; 593 entry.lengths[i] = 0; 669 private final long[] lengths; field in class:DiskLruCache.Value 672 private Value(String key, long sequenceNumber, File[] files, long[] lengths) { argument 676 this.lengths 807 private final long[] lengths; field in class:DiskLruCache.Entry [all...] |
/external/icu/icu4c/source/test/perf/dicttrieperf/ |
H A D | dicttrieperf.cpp | 218 // Prime the prefix lengths so that we don't keep prefixLength at 0 until 351 int32_t *lengths, int &count, int limit ) { 367 // lengths[count++]=(int32_t)utext_getNativeIndex(text); 368 lengths[count++]=numChars; // CompactTrieDictionary just counts chars too. 444 int32_t lengths[20]; local 455 lengths, count, UPRV_LENGTHOF(lengths)); 456 if(count==0 || lengths[count-1]!=lines[i].len) { 550 int32_t *lengths, int &count, int limit ) { 561 // lengths[coun 349 ucharsTrieMatches(UCharsTrie &trie, UText *text, int32_t textLimit, int32_t *lengths, int &count, int limit ) argument 548 bytesTrieMatches(BytesTrie &trie, UText *text, int32_t textLimit, int32_t *lengths, int &count, int limit ) argument 593 int32_t lengths[20]; local [all...] |
/external/tensorflow/tensorflow/compiler/tf2xla/kernels/ |
H A D | tensor_array_ops.cc | 463 Tensor lengths(DT_INT64, {ta_dims[0]}); 464 auto lengths_vec = lengths.vec<int64>(); 468 ctx->SetConstantOutput(1, lengths); 486 std::vector<int64> lengths; variable 487 OP_REQUIRES_OK(ctx, ctx->ConstantInputAsIntVector(2, &lengths)); 490 if (!lengths.empty()) { 491 length = lengths[0]; 492 for (int i = 1; i < lengths.size(); ++i) { 493 OP_REQUIRES(ctx, lengths[i] == length, 494 errors::InvalidArgument("lengths mus [all...] |
/external/tensorflow/tensorflow/python/keras/_impl/keras/preprocessing/ |
H A D | sequence.py | 62 lengths = [] 67 lengths.append(len(x)) 71 maxlen = np.max(lengths)
|
/external/tensorflow/tensorflow/python/ops/ |
H A D | tensor_array_grad.py | 211 lengths = op.outputs[1] 217 u_g = g.split(grad, lengths=lengths) 242 # handle, value, lengths, flow_in
|
H A D | tensor_array_ops.py | 359 def split(self, value, lengths, name=None): 362 [self._handle, value, lengths]): 365 lengths_64 = math_ops.to_int64(lengths) 376 lengths=lengths_64, 669 def split(self, value, lengths, name=None): 673 lengths = constant_op.constant(lengths) 674 sum_lengths = math_ops.reduce_sum(lengths) 675 if lengths.shape.ndims != 1: 677 None, None, "Expected lengths t [all...] |
/external/okhttp/okhttp/src/main/java/com/squareup/okhttp/internal/ |
H A D | DiskLruCache.java | 130 * and may be read. A publish line is followed by the lengths of each of 360 size += entry.lengths[t]; 540 long oldLength = entry.lengths[i]; 542 entry.lengths[i] = newLength; 608 size -= entry.lengths[i]; 609 entry.lengths[i] = 0; 770 private final long[] lengths; field in class:DiskLruCache.Snapshot 772 private Snapshot(String key, long sequenceNumber, Source[] sources, long[] lengths) { argument 776 this.lengths = lengths; 958 private final long[] lengths; field in class:DiskLruCache.Entry [all...] |
/external/mesa3d/src/mesa/math/ |
H A D | m_xform.h | 117 const GLfloat lengths[],
|
H A D | m_debug_norm.c | 127 const GLfloat *lengths, 135 (void) lengths; 150 const GLfloat *lengths, 163 if ( !lengths ) { 175 scale = lengths[i]; 124 ref_norm_transform_rescale( const GLmatrix *mat, GLfloat scale, const GLvector4f *in, const GLfloat *lengths, GLvector4f *dest ) argument 147 ref_norm_transform_normalize( const GLmatrix *mat, GLfloat scale, const GLvector4f *in, const GLfloat *lengths, GLvector4f *dest ) argument
|
/external/tensorflow/tensorflow/contrib/eager/python/examples/spinn/ |
H A D | data.py | 204 """Calculate bin boundaries given a histogram of lengths and minimum bin size. 218 lengths = sorted(length2count.keys()) 220 for length in lengths: 225 if bounds[-1] != lengths[-1]: 226 bounds.append(lengths[-1]) 349 # lengths in the pair. During training, the batches are expected to be 359 # similar max lengths are batched together, minimizing the inefficiency 360 # due to uneven max lengths. The sentences are batched differently in 363 # take care of any remaining unevenness of the max sentence lengths.
|
/external/mesa3d/src/gallium/state_trackers/clover/api/ |
H A D | program.cpp | 52 const char **strings, const size_t *lengths, 63 source += (lengths && lengths[i] ? 64 std::string(strings[i], strings[i] + lengths[i]) : 79 const size_t *lengths, 85 if (!lengths || !binaries) 110 range(lengths, n)); 51 clCreateProgramWithSource(cl_context d_ctx, cl_uint count, const char **strings, const size_t *lengths, cl_int *r_errcode) argument 77 clCreateProgramWithBinary(cl_context d_ctx, cl_uint n, const cl_device_id *d_devs, const size_t *lengths, const unsigned char **binaries, cl_int *r_status, cl_int *r_errcode) argument
|