/external/guava/guava/src/com/google/common/cache/ |
H A D | LocalCache.java | 110 * This implementation uses a per-segment queue to record a memento of the additions, removals, 111 * and accesses that were performed on the map. The queue is drained on writes and when it exceeds 146 * the cleanup queue and both reference queues. 651 ValueReference<K, V> copyFor(ReferenceQueue<V> queue, ReferenceEntry<K, V> entry); argument 697 ReferenceQueue<Object> queue, ReferenceEntry<Object, Object> entry) { 787 * Returns the next entry in the access queue. 792 * Sets the next entry in the access queue. 797 * Returns the previous entry in the access queue. 802 * Sets the previous entry in the access queue. 823 * Returns the next entry in the write queue 1346 WeakEntry(ReferenceQueue<K> queue, K key, int hash, @Nullable ReferenceEntry<K, V> next) argument 1450 WeakAccessEntry( ReferenceQueue<K> queue, K key, int hash, @Nullable ReferenceEntry<K, V> next) argument 1498 WeakWriteEntry( ReferenceQueue<K> queue, K key, int hash, @Nullable ReferenceEntry<K, V> next) argument 1546 WeakAccessWriteEntry( ReferenceQueue<K> queue, K key, int hash, @Nullable ReferenceEntry<K, V> next) argument 1639 WeakValueReference(ReferenceQueue<V> queue, V referent, ReferenceEntry<K, V> entry) argument 1658 copyFor( ReferenceQueue<V> queue, ReferenceEntry<K, V> entry) argument 1686 SoftValueReference(ReferenceQueue<V> queue, V referent, ReferenceEntry<K, V> entry) argument 1705 copyFor(ReferenceQueue<V> queue, ReferenceEntry<K, V> entry) argument 1751 copyFor(ReferenceQueue<V> queue, ReferenceEntry<K, V> entry) argument 1780 WeightedWeakValueReference(ReferenceQueue<V> queue, V referent, ReferenceEntry<K, V> entry, int weight) argument 1792 copyFor( ReferenceQueue<V> queue, ReferenceEntry<K, V> entry) argument 1804 WeightedSoftValueReference(ReferenceQueue<V> queue, V referent, ReferenceEntry<K, V> entry, int weight) argument 1815 copyFor(ReferenceQueue<V> queue, ReferenceEntry<K, V> entry) argument 3604 copyFor(ReferenceQueue<V> queue, ReferenceEntry<K, V> entry) argument [all...] |
/external/guava/guava/src/com/google/common/collect/ |
H A D | ComputingConcurrentHashMap.java | 117 // entries, but let's accomodate an incorrect expiration queue. 226 public ValueReference<K, V> copyFor(ReferenceQueue<V> queue, ReferenceEntry<K, V> entry) { argument 265 public ValueReference<K, V> copyFor(ReferenceQueue<V> queue, ReferenceEntry<K, V> entry) { argument 306 public ValueReference<K, V> copyFor(ReferenceQueue<V> queue, ReferenceEntry<K, V> entry) { argument
|
H A D | Iterables.java | 972 * queue's elements in queue order 998 private final Queue<T> queue; field in class:Iterables.ConsumingQueueIterator 1000 private ConsumingQueueIterator(Queue<T> queue) { argument 1001 this.queue = queue; 1006 return queue.remove();
|
H A D | Iterators.java | 1287 final Queue<PeekingIterator<T>> queue; field in class:Iterators.MergingIterator 1304 queue = new PriorityQueue<PeekingIterator<T>>(2, heapComparator); 1308 queue.add(Iterators.peekingIterator(iterator)); 1315 if (queue.isEmpty()) { 1319 PeekingIterator<T> nextIter = queue.poll(); 1323 queue.add(nextIter);
|
H A D | MapMakerInternalMap.java | 91 * This implementation uses a per-segment queue to record a memento of the additions, removals, 92 * and accesses that were performed on the map. The queue is drained on writes and when it exceeds 127 * the cleanup queue and both reference queues. 609 ValueReference<K, V> copyFor(ReferenceQueue<V> queue, ReferenceEntry<K, V> entry); argument 643 ReferenceQueue<Object> queue, ReferenceEntry<Object, Object> entry) { 1224 SoftEntry(ReferenceQueue<K> queue, K key, int hash, @Nullable ReferenceEntry<K, V> next) { argument 1225 super(key, queue); 1320 ReferenceQueue<K> queue, K key, int hash, @Nullable ReferenceEntry<K, V> next) { 1321 super(queue, key, hash, next); 1368 ReferenceQueue<K> queue, 1319 SoftExpirableEntry( ReferenceQueue<K> queue, K key, int hash, @Nullable ReferenceEntry<K, V> next) argument 1367 SoftEvictableEntry( ReferenceQueue<K> queue, K key, int hash, @Nullable ReferenceEntry<K, V> next) argument 1403 SoftExpirableEvictableEntry( ReferenceQueue<K> queue, K key, int hash, @Nullable ReferenceEntry<K, V> next) argument 1481 WeakEntry(ReferenceQueue<K> queue, K key, int hash, @Nullable ReferenceEntry<K, V> next) argument 1577 WeakExpirableEntry( ReferenceQueue<K> queue, K key, int hash, @Nullable ReferenceEntry<K, V> next) argument 1625 WeakEvictableEntry( ReferenceQueue<K> queue, K key, int hash, @Nullable ReferenceEntry<K, V> next) argument 1661 WeakExpirableEvictableEntry( ReferenceQueue<K> queue, K key, int hash, @Nullable ReferenceEntry<K, V> next) argument 1742 WeakValueReference(ReferenceQueue<V> queue, V referent, ReferenceEntry<K, V> entry) argument 1758 copyFor( ReferenceQueue<V> queue, ReferenceEntry<K, V> entry) argument 1781 SoftValueReference(ReferenceQueue<V> queue, V referent, ReferenceEntry<K, V> entry) argument 1797 copyFor(ReferenceQueue<V> queue, ReferenceEntry<K, V> entry) argument 1833 copyFor(ReferenceQueue<V> queue, ReferenceEntry<K, V> entry) argument [all...] |
H A D | MinMaxPriorityQueue.java | 42 * A double-ended priority queue, which provides constant-time access to both 43 * its least element and its greatest element, as determined by the queue's 50 * the queue according to the queue's comparator. But unlike a regular priority 51 * queue, the methods {@link #peekLast}, {@link #pollLast} and 53 * in the queue instead. 55 * <p>A min-max priority queue can be configured with a maximum size. If so, 56 * each time the size of the queue exceeds that value, the queue automatically 79 * <li>If you only access one end of the queue, an 222 private Object[] queue; field in class:MinMaxPriorityQueue [all...] |
/external/guava/guava/src/com/google/common/util/concurrent/ |
H A D | Uninterruptibles.java | 207 * Invokes {@code queue.}{@link BlockingQueue#take() take()} uninterruptibly. 209 public static <E> E takeUninterruptibly(BlockingQueue<E> queue) { argument 214 return queue.take(); 227 * Invokes {@code queue.}{@link BlockingQueue#put(Object) put(element)} 230 public static <E> void putUninterruptibly(BlockingQueue<E> queue, E element) { argument 235 queue.put(element);
|
/external/iproute2/ip/ |
H A D | ipntable.c | 54 " [ delay_probe MSEC ] [ queue LEN ]\n" 204 } else if (strcmp(*argv, "queue") == 0) { 205 __u32 queue; local 209 if (get_u32(&queue, *argv, 0)) 210 invarg("\"queue\" value is invalid", *argv); 215 NDTPA_QUEUE_LEN, queue); 511 __u32 queue = rta_getattr_u32(tpb[NDTPA_QUEUE_LEN]); local 512 fprintf(fp, "queue %u ", queue);
|
/external/javassist/src/main/javassist/bytecode/analysis/ |
H A D | Analyzer.java | 133 IntQueue queue = new IntQueue(); 141 queue.add(iter.next()); 142 while (!queue.isEmpty()) { 143 analyzeNextEntry(method, iter, queue, executor); 168 IntQueue queue, Executor executor) throws BadBytecode { 169 int pos = queue.take(); 185 mergeTableSwitch(queue, pos, iter, frame); 187 mergeLookupSwitch(queue, pos, iter, frame); 189 mergeRet(queue, iter, pos, frame, subroutine); 195 mergeJsr(queue, frame 167 analyzeNextEntry(MethodInfo method, CodeIterator iter, IntQueue queue, Executor executor) argument 276 merge(IntQueue queue, Frame frame, int target) argument 292 mergeExceptionHandlers(IntQueue queue, MethodInfo method, int pos, Frame frame) argument 306 mergeJsr(IntQueue queue, Frame frame, Subroutine sub, int pos, int next) argument 347 mergeLookupSwitch(IntQueue queue, int pos, CodeIterator iter, Frame frame) argument 361 mergeRet(IntQueue queue, CodeIterator iter, int pos, Frame frame, Subroutine subroutine) argument 400 mergeTableSwitch(IntQueue queue, int pos, CodeIterator iter, Frame frame) argument [all...] |
/external/javassist/src/main/javassist/scopedpool/ |
H A D | SoftValueHashMap.java | 62 /* Reference queue for cleared WeakKeys */ 63 private ReferenceQueue queue = new ReferenceQueue(); field in class:SoftValueHashMap 71 while ((ref = (SoftValueRef)queue.poll()) != null) { 204 Object rtn = hash.put(key, SoftValueRef.create(key, value, queue));
|
/external/kernel-headers/original/uapi/sound/ |
H A D | asequencer.h | 73 #define SNDRV_SEQ_EVENT_SETPOS_TICK 33 /* set tick queue position */ 74 #define SNDRV_SEQ_EVENT_SETPOS_TIME 34 /* set realtime queue position */ 78 #define SNDRV_SEQ_EVENT_QUEUE_SKEW 38 /* skew queue tempo */ 248 /* queue timer control */ 250 unsigned char queue; /* affected queue */ member in struct:snd_seq_ev_queue_control 276 unsigned char queue; /* schedule queue */ member in struct:snd_seq_event 289 struct snd_seq_ev_queue_control queue; member in union:snd_seq_event::__anon23538 393 unsigned char queue; /* Queu member in struct:snd_seq_remove_events 472 int queue; /* queue id */ member in struct:snd_seq_queue_info 488 int queue; /* queue id */ member in struct:snd_seq_queue_status 500 int queue; /* sequencer queue */ member in struct:snd_seq_queue_tempo 516 int queue; /* sequencer queue */ member in struct:snd_seq_queue_timer 529 int queue; /* sequencer queue */ member in struct:snd_seq_queue_client 547 unsigned char queue; /* input time-stamp queue (optional) */ member in struct:snd_seq_port_subscribe 562 unsigned char queue; /* R/O: result */ member in struct:snd_seq_query_subs [all...] |
H A D | asound.h | 583 size_t buffer_size; /* queue size in bytes */ 691 #define SNDRV_TIMER_PSFLG_EARLY_EVENT (1<<2) /* write early event to the poll queue */ 696 unsigned int queue_size; /* total size of queue (32-1024) */ 706 unsigned int overrun; /* count of read queue overruns */ 707 unsigned int queue; /* used queue size */ member in struct:snd_timer_status
|
/external/libnl/lib/netfilter/ |
H A D | queue.c | 2 * lib/netfilter/queue.c Netfilter Queue 14 * @defgroup queue Queue 25 #include <netlink/netfilter/queue.h> 111 static int nfnl_queue_build_request(const struct nfnl_queue *queue, argument 116 if (!nfnl_queue_test_group(queue)) 120 0, nfnl_queue_get_group(queue)); 124 if (nfnl_queue_test_maxlen(queue) && 126 htonl(nfnl_queue_get_maxlen(queue))) < 0) 132 if (nfnl_queue_test_copy_mode(queue)) { 135 switch (nfnl_queue_get_copy_mode(queue)) { 160 nfnl_queue_build_create_request(const struct nfnl_queue *queue, struct nl_msg **result) argument 182 nfnl_queue_create(struct nl_sock *nlh, const struct nfnl_queue *queue) argument 193 nfnl_queue_build_change_request(const struct nfnl_queue *queue, struct nl_msg **result) argument 199 nfnl_queue_change(struct nl_sock *nlh, const struct nfnl_queue *queue) argument 210 nfnl_queue_build_delete_request(const struct nfnl_queue *queue, struct nl_msg **result) argument 220 nfnl_queue_delete(struct nl_sock *nlh, const struct nfnl_queue *queue) argument [all...] |
H A D | queue_obj.c | 14 * @defgroup queue Queue 21 #include <netlink/netfilter/queue.h> 33 struct nfnl_queue *queue = (struct nfnl_queue *) a; local 38 if (queue->ce_mask & QUEUE_ATTR_GROUP) 39 nl_dump(p, "group=%u ", queue->queue_group); 41 if (queue->ce_mask & QUEUE_ATTR_MAXLEN) 42 nl_dump(p, "maxlen=%u ", queue->queue_maxlen); 44 if (queue->ce_mask & QUEUE_ATTR_COPY_MODE) 46 nfnl_queue_copy_mode2str(queue->queue_copy_mode, 49 if (queue 83 nfnl_queue_get(struct nfnl_queue *queue) argument 88 nfnl_queue_put(struct nfnl_queue *queue) argument 100 nfnl_queue_set_group(struct nfnl_queue *queue, uint16_t group) argument 106 nfnl_queue_test_group(const struct nfnl_queue *queue) argument 111 nfnl_queue_get_group(const struct nfnl_queue *queue) argument 116 nfnl_queue_set_maxlen(struct nfnl_queue *queue, uint32_t maxlen) argument 122 nfnl_queue_test_maxlen(const struct nfnl_queue *queue) argument 127 nfnl_queue_get_maxlen(const struct nfnl_queue *queue) argument 132 nfnl_queue_set_copy_mode(struct nfnl_queue *queue, enum nfnl_queue_copy_mode mode) argument 138 nfnl_queue_test_copy_mode(const struct nfnl_queue *queue) argument 143 nfnl_queue_get_copy_mode(const struct nfnl_queue *queue) argument 148 nfnl_queue_set_copy_range(struct nfnl_queue *queue, uint32_t copy_range) argument 154 nfnl_queue_test_copy_range(const struct nfnl_queue *queue) argument 159 nfnl_queue_get_copy_range(const struct nfnl_queue *queue) argument [all...] |
/external/libnl/src/ |
H A D | nf-queue.c | 2 * src/nf-queue.c Monitor netfilter queue events 19 #include <netlink/netfilter/queue.h> 26 struct nfnl_queue *queue; local 28 queue = nfnl_queue_alloc(); 29 if (!queue) 30 nl_cli_fatal(ENOMEM, "Unable to allocate queue object"); 32 return queue; 87 struct nfnl_queue *queue; local 101 printf("Usage: nf-queue famil [all...] |
/external/linux-tools-perf/perf-3.12.0/tools/perf/util/ |
H A D | annotate.c | 642 int max_lines, struct disasm_line *queue) 685 if (queue != NULL) { 686 list_for_each_entry_from(queue, ¬es->src->source, node) { 687 if (queue == dl) 689 disasm_line__print(queue, sym, start, evsel, len, 728 if (queue) 1222 struct disasm_line *pos, *queue = NULL; local 1254 if (context && queue == NULL) { 1255 queue = pos; 1261 queue)) { 640 disasm_line__print(struct disasm_line *dl, struct symbol *sym, u64 start, struct perf_evsel *evsel, u64 len, int min_pcnt, int printed, int max_lines, struct disasm_line *queue) argument [all...] |
/external/lldb/source/Plugins/OperatingSystem/Python/ |
H A D | OperatingSystemPython.cpp | 241 PythonString queue_pystr("queue"); 249 const char *queue = thread_dict.GetItemForKeyAsString (queue_pystr); local 275 queue,
|
/external/lldb/source/Plugins/Process/Utility/ |
H A D | ThreadMemory.cpp | 36 const char *queue, 47 if (queue) 48 m_queue = queue; 33 ThreadMemory(Process &process, lldb::tid_t tid, const char *name, const char *queue, lldb::addr_t register_data_addr) argument
|
/external/llvm/utils/lit/lit/ |
H A D | run.py | 6 import Queue as queue namespace 8 import queue namespace 50 # Create a shared queue to provide the test indices. 51 self.queue = queue_impl() 53 self.queue.put(i) 55 self.queue.put(None) 66 return self.queue.get() 116 self.queue = multiprocessing.Queue() 120 # results to the actual display implementation via an output queue. 121 self.queue [all...] |
/external/mesa3d/include/CL/ |
H A D | cl.hpp | 113 * cl::CommandQueue queue(context, devices[0], 0, &err); 114 * queue.enqueueNDRangeKernel( 1038 static cl_int retain(cl_command_queue queue) argument 1039 { return ::clRetainCommandQueue(queue); } 1040 static cl_int release(cl_command_queue queue) argument 1041 { return ::clReleaseCommandQueue(queue); } 2370 const CommandQueue& queue, 2376 const CommandQueue& queue, 3140 const CommandQueue& queue, 3145 queue_(queue), 3138 KernelFunctor( const Kernel& kernel, const CommandQueue& queue, const NDRange& offset, const NDRange& global, const NDRange& local) argument 3355 bind( const CommandQueue& queue, const NDRange& offset, const NDRange& global, const NDRange& local) argument 3364 bind( const CommandQueue& queue, const NDRange& global, const NDRange& local) argument [all...] |
/external/mesa3d/src/gallium/drivers/llvmpipe/ |
H A D | lp_scene_queue.c | 30 * Scene queue. We'll use two queues. One contains "full" scenes which 49 * A queue of scenes 58 /** Allocate a new scene queue */ 62 struct lp_scene_queue *queue = CALLOC_STRUCT(lp_scene_queue); local 63 if (queue == NULL) 66 queue->ring = util_ringbuffer_create( MAX_SCENE_QUEUE * 68 if (queue->ring == NULL) 71 return queue; 74 FREE(queue); 79 /** Delete a scene queue */ 81 lp_scene_queue_destroy(struct lp_scene_queue *queue) argument 90 lp_scene_dequeue(struct lp_scene_queue *queue, boolean wait) argument 110 lp_scene_enqueue(struct lp_scene_queue *queue, struct lp_scene *scene) argument [all...] |
/external/mesa3d/src/gallium/drivers/svga/ |
H A D | svga_state_rss.c | 48 svga_queue_rs( &queue, SVGA3D_RS_##token, value ); \ 58 svga_queue_rs( &queue, SVGA3D_RS_##token, value ); \ 83 struct rs_queue queue; local 86 queue.rs_count = 0; 268 if (queue.rs_count) { 273 queue.rs_count ) != PIPE_OK) 277 queue.rs, 278 queue.rs_count * sizeof queue.rs[0]);
|
H A D | svga_state_tss.c | 77 struct bind_queue queue; local 79 queue.bind_count = 0; 123 queue.bind[queue.bind_count].unit = i; 124 queue.bind[queue.bind_count].view = view; 125 queue.bind_count++; 134 if (queue.bind_count) { 139 queue.bind_count ) != PIPE_OK) 142 for (i = 0; i < queue 187 struct bind_queue queue; local 294 struct ts_queue queue; local [all...] |
/external/mesa3d/src/gallium/state_trackers/clover/core/ |
H A D | event.cpp | 86 pipe_screen *screen = queue()->dev.pipe; 92 pipe_screen *screen = queue()->dev.pipe; 108 hard_event::queue() const { function in class:hard_event 119 pipe_screen *screen = queue()->dev.pipe; 122 queue()->flush(); 131 pipe_screen *screen = queue()->dev.pipe; 159 soft_event::queue() const { function in class:soft_event
|
/external/mesa3d/src/gallium/state_trackers/egl/wayland/ |
H A D | native_wayland.h | 48 struct wl_event_queue *queue; member in struct:wayland_display
|