Lines Matching refs:beg

117   ShadowSegmentEndpoint beg(beg_addr);
119 if (beg.chunk == end.chunk) {
120 CHECK(beg.offset < end.offset);
121 s8 value = beg.value;
126 if (beg.offset > 0) {
127 *beg.chunk = Min(value, beg.offset);
129 *beg.chunk = kAsanUserPoisonedMemoryMagic;
134 CHECK(beg.chunk < end.chunk);
135 if (beg.offset > 0) {
136 // Mark bytes from beg.offset as unaddressable.
137 if (beg.value == 0) {
138 *beg.chunk = beg.offset;
140 *beg.chunk = Min(beg.value, beg.offset);
142 beg.chunk++;
144 REAL(memset)(beg.chunk, kAsanUserPoisonedMemoryMagic, end.chunk - beg.chunk);
157 ShadowSegmentEndpoint beg(beg_addr);
159 if (beg.chunk == end.chunk) {
160 CHECK(beg.offset < end.offset);
161 s8 value = beg.value;
166 *beg.chunk = Max(value, end.offset);
170 CHECK(beg.chunk < end.chunk);
171 if (beg.offset > 0) {
172 *beg.chunk = 0;
173 beg.chunk++;
175 REAL(memset)(beg.chunk, 0, end.chunk - beg.chunk);
185 uptr __asan_region_is_poisoned(uptr beg, uptr size) {
187 uptr end = beg + size;
188 if (!AddrIsInMem(beg)) return beg;
190 CHECK_LT(beg, end);
191 uptr aligned_b = RoundUpTo(beg, SHADOW_GRANULARITY);
198 if (!__asan::AddressIsPoisoned(beg) &&
206 for (; beg < end; beg++)
207 if (__asan::AddressIsPoisoned(beg))
208 return beg;
334 uptr beg = reinterpret_cast<uptr>(beg_p);
339 if (!(beg <= old_mid && beg <= new_mid && old_mid <= end && new_mid <= end &&
340 IsAligned(beg, granularity))) {
342 ReportBadParamsToAnnotateContiguousContainer(beg, end, old_mid, new_mid,
345 CHECK_LE(end - beg,
382 uptr beg = reinterpret_cast<uptr>(beg_p);
385 CHECK_LE(beg, mid);
387 // Check some bytes starting from beg, some bytes around mid, and some bytes
390 uptr r1_beg = beg;
392 uptr r2_beg = Max(beg, mid - kMaxRangeToCheck);