Parcel.cpp revision bca20d8376708da7ed835e71442ab52197d1e9c0
1/* 2 * Copyright (C) 2005 The Android Open Source Project 3 * 4 * Licensed under the Apache License, Version 2.0 (the "License"); 5 * you may not use this file except in compliance with the License. 6 * You may obtain a copy of the License at 7 * 8 * http://www.apache.org/licenses/LICENSE-2.0 9 * 10 * Unless required by applicable law or agreed to in writing, software 11 * distributed under the License is distributed on an "AS IS" BASIS, 12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 13 * See the License for the specific language governing permissions and 14 * limitations under the License. 15 */ 16 17#define LOG_TAG "Parcel" 18//#define LOG_NDEBUG 0 19 20#include <binder/Parcel.h> 21 22#include <binder/IPCThreadState.h> 23#include <binder/Binder.h> 24#include <binder/BpBinder.h> 25#include <binder/ProcessState.h> 26#include <binder/TextOutput.h> 27 28#include <errno.h> 29#include <utils/Debug.h> 30#include <utils/Log.h> 31#include <utils/String8.h> 32#include <utils/String16.h> 33#include <utils/misc.h> 34#include <utils/Flattenable.h> 35#include <cutils/ashmem.h> 36 37#include <private/binder/binder_module.h> 38#include <private/binder/Static.h> 39 40#include <inttypes.h> 41#include <stdio.h> 42#include <stdlib.h> 43#include <stdint.h> 44#include <sys/mman.h> 45 46#ifndef INT32_MAX 47#define INT32_MAX ((int32_t)(2147483647)) 48#endif 49 50#define LOG_REFS(...) 51//#define LOG_REFS(...) ALOG(LOG_DEBUG, "Parcel", __VA_ARGS__) 52#define LOG_ALLOC(...) 53//#define LOG_ALLOC(...) ALOG(LOG_DEBUG, "Parcel", __VA_ARGS__) 54 55// --------------------------------------------------------------------------- 56 57// This macro should never be used at runtime, as a too large value 58// of s could cause an integer overflow. Instead, you should always 59// use the wrapper function pad_size() 60#define PAD_SIZE_UNSAFE(s) (((s)+3)&~3) 61 62static size_t pad_size(size_t s) { 63 if (s > (SIZE_T_MAX - 3)) { 64 abort(); 65 } 66 return PAD_SIZE_UNSAFE(s); 67} 68 69// Note: must be kept in sync with android/os/StrictMode.java's PENALTY_GATHER 70#define STRICT_MODE_PENALTY_GATHER (0x40 << 16) 71 72// Note: must be kept in sync with android/os/Parcel.java's EX_HAS_REPLY_HEADER 73#define EX_HAS_REPLY_HEADER -128 74 75// XXX This can be made public if we want to provide 76// support for typed data. 77struct small_flat_data 78{ 79 uint32_t type; 80 uint32_t data; 81}; 82 83namespace android { 84 85static pthread_mutex_t gParcelGlobalAllocSizeLock = PTHREAD_MUTEX_INITIALIZER; 86static size_t gParcelGlobalAllocSize = 0; 87static size_t gParcelGlobalAllocCount = 0; 88 89// Maximum size of a blob to transfer in-place. 90static const size_t BLOB_INPLACE_LIMIT = 16 * 1024; 91 92enum { 93 BLOB_INPLACE = 0, 94 BLOB_ASHMEM_IMMUTABLE = 1, 95 BLOB_ASHMEM_MUTABLE = 2, 96}; 97 98void acquire_object(const sp<ProcessState>& proc, 99 const flat_binder_object& obj, const void* who) 100{ 101 switch (obj.type) { 102 case BINDER_TYPE_BINDER: 103 if (obj.binder) { 104 LOG_REFS("Parcel %p acquiring reference on local %p", who, obj.cookie); 105 reinterpret_cast<IBinder*>(obj.cookie)->incStrong(who); 106 } 107 return; 108 case BINDER_TYPE_WEAK_BINDER: 109 if (obj.binder) 110 reinterpret_cast<RefBase::weakref_type*>(obj.binder)->incWeak(who); 111 return; 112 case BINDER_TYPE_HANDLE: { 113 const sp<IBinder> b = proc->getStrongProxyForHandle(obj.handle); 114 if (b != NULL) { 115 LOG_REFS("Parcel %p acquiring reference on remote %p", who, b.get()); 116 b->incStrong(who); 117 } 118 return; 119 } 120 case BINDER_TYPE_WEAK_HANDLE: { 121 const wp<IBinder> b = proc->getWeakProxyForHandle(obj.handle); 122 if (b != NULL) b.get_refs()->incWeak(who); 123 return; 124 } 125 case BINDER_TYPE_FD: { 126 // intentionally blank -- nothing to do to acquire this, but we do 127 // recognize it as a legitimate object type. 128 return; 129 } 130 } 131 132 ALOGD("Invalid object type 0x%08x", obj.type); 133} 134 135void release_object(const sp<ProcessState>& proc, 136 const flat_binder_object& obj, const void* who) 137{ 138 switch (obj.type) { 139 case BINDER_TYPE_BINDER: 140 if (obj.binder) { 141 LOG_REFS("Parcel %p releasing reference on local %p", who, obj.cookie); 142 reinterpret_cast<IBinder*>(obj.cookie)->decStrong(who); 143 } 144 return; 145 case BINDER_TYPE_WEAK_BINDER: 146 if (obj.binder) 147 reinterpret_cast<RefBase::weakref_type*>(obj.binder)->decWeak(who); 148 return; 149 case BINDER_TYPE_HANDLE: { 150 const sp<IBinder> b = proc->getStrongProxyForHandle(obj.handle); 151 if (b != NULL) { 152 LOG_REFS("Parcel %p releasing reference on remote %p", who, b.get()); 153 b->decStrong(who); 154 } 155 return; 156 } 157 case BINDER_TYPE_WEAK_HANDLE: { 158 const wp<IBinder> b = proc->getWeakProxyForHandle(obj.handle); 159 if (b != NULL) b.get_refs()->decWeak(who); 160 return; 161 } 162 case BINDER_TYPE_FD: { 163 if (obj.cookie != 0) close(obj.handle); 164 return; 165 } 166 } 167 168 ALOGE("Invalid object type 0x%08x", obj.type); 169} 170 171inline static status_t finish_flatten_binder( 172 const sp<IBinder>& /*binder*/, const flat_binder_object& flat, Parcel* out) 173{ 174 return out->writeObject(flat, false); 175} 176 177status_t flatten_binder(const sp<ProcessState>& /*proc*/, 178 const sp<IBinder>& binder, Parcel* out) 179{ 180 flat_binder_object obj; 181 182 obj.flags = 0x7f | FLAT_BINDER_FLAG_ACCEPTS_FDS; 183 if (binder != NULL) { 184 IBinder *local = binder->localBinder(); 185 if (!local) { 186 BpBinder *proxy = binder->remoteBinder(); 187 if (proxy == NULL) { 188 ALOGE("null proxy"); 189 } 190 const int32_t handle = proxy ? proxy->handle() : 0; 191 obj.type = BINDER_TYPE_HANDLE; 192 obj.binder = 0; /* Don't pass uninitialized stack data to a remote process */ 193 obj.handle = handle; 194 obj.cookie = 0; 195 } else { 196 obj.type = BINDER_TYPE_BINDER; 197 obj.binder = reinterpret_cast<uintptr_t>(local->getWeakRefs()); 198 obj.cookie = reinterpret_cast<uintptr_t>(local); 199 } 200 } else { 201 obj.type = BINDER_TYPE_BINDER; 202 obj.binder = 0; 203 obj.cookie = 0; 204 } 205 206 return finish_flatten_binder(binder, obj, out); 207} 208 209status_t flatten_binder(const sp<ProcessState>& /*proc*/, 210 const wp<IBinder>& binder, Parcel* out) 211{ 212 flat_binder_object obj; 213 214 obj.flags = 0x7f | FLAT_BINDER_FLAG_ACCEPTS_FDS; 215 if (binder != NULL) { 216 sp<IBinder> real = binder.promote(); 217 if (real != NULL) { 218 IBinder *local = real->localBinder(); 219 if (!local) { 220 BpBinder *proxy = real->remoteBinder(); 221 if (proxy == NULL) { 222 ALOGE("null proxy"); 223 } 224 const int32_t handle = proxy ? proxy->handle() : 0; 225 obj.type = BINDER_TYPE_WEAK_HANDLE; 226 obj.binder = 0; /* Don't pass uninitialized stack data to a remote process */ 227 obj.handle = handle; 228 obj.cookie = 0; 229 } else { 230 obj.type = BINDER_TYPE_WEAK_BINDER; 231 obj.binder = reinterpret_cast<uintptr_t>(binder.get_refs()); 232 obj.cookie = reinterpret_cast<uintptr_t>(binder.unsafe_get()); 233 } 234 return finish_flatten_binder(real, obj, out); 235 } 236 237 // XXX How to deal? In order to flatten the given binder, 238 // we need to probe it for information, which requires a primary 239 // reference... but we don't have one. 240 // 241 // The OpenBinder implementation uses a dynamic_cast<> here, 242 // but we can't do that with the different reference counting 243 // implementation we are using. 244 ALOGE("Unable to unflatten Binder weak reference!"); 245 obj.type = BINDER_TYPE_BINDER; 246 obj.binder = 0; 247 obj.cookie = 0; 248 return finish_flatten_binder(NULL, obj, out); 249 250 } else { 251 obj.type = BINDER_TYPE_BINDER; 252 obj.binder = 0; 253 obj.cookie = 0; 254 return finish_flatten_binder(NULL, obj, out); 255 } 256} 257 258inline static status_t finish_unflatten_binder( 259 BpBinder* /*proxy*/, const flat_binder_object& /*flat*/, 260 const Parcel& /*in*/) 261{ 262 return NO_ERROR; 263} 264 265status_t unflatten_binder(const sp<ProcessState>& proc, 266 const Parcel& in, sp<IBinder>* out) 267{ 268 const flat_binder_object* flat = in.readObject(false); 269 270 if (flat) { 271 switch (flat->type) { 272 case BINDER_TYPE_BINDER: 273 *out = reinterpret_cast<IBinder*>(flat->cookie); 274 return finish_unflatten_binder(NULL, *flat, in); 275 case BINDER_TYPE_HANDLE: 276 *out = proc->getStrongProxyForHandle(flat->handle); 277 return finish_unflatten_binder( 278 static_cast<BpBinder*>(out->get()), *flat, in); 279 } 280 } 281 return BAD_TYPE; 282} 283 284status_t unflatten_binder(const sp<ProcessState>& proc, 285 const Parcel& in, wp<IBinder>* out) 286{ 287 const flat_binder_object* flat = in.readObject(false); 288 289 if (flat) { 290 switch (flat->type) { 291 case BINDER_TYPE_BINDER: 292 *out = reinterpret_cast<IBinder*>(flat->cookie); 293 return finish_unflatten_binder(NULL, *flat, in); 294 case BINDER_TYPE_WEAK_BINDER: 295 if (flat->binder != 0) { 296 out->set_object_and_refs( 297 reinterpret_cast<IBinder*>(flat->cookie), 298 reinterpret_cast<RefBase::weakref_type*>(flat->binder)); 299 } else { 300 *out = NULL; 301 } 302 return finish_unflatten_binder(NULL, *flat, in); 303 case BINDER_TYPE_HANDLE: 304 case BINDER_TYPE_WEAK_HANDLE: 305 *out = proc->getWeakProxyForHandle(flat->handle); 306 return finish_unflatten_binder( 307 static_cast<BpBinder*>(out->unsafe_get()), *flat, in); 308 } 309 } 310 return BAD_TYPE; 311} 312 313// --------------------------------------------------------------------------- 314 315Parcel::Parcel() 316{ 317 LOG_ALLOC("Parcel %p: constructing", this); 318 initState(); 319} 320 321Parcel::~Parcel() 322{ 323 freeDataNoInit(); 324 LOG_ALLOC("Parcel %p: destroyed", this); 325} 326 327size_t Parcel::getGlobalAllocSize() { 328 pthread_mutex_lock(&gParcelGlobalAllocSizeLock); 329 size_t size = gParcelGlobalAllocSize; 330 pthread_mutex_unlock(&gParcelGlobalAllocSizeLock); 331 return size; 332} 333 334size_t Parcel::getGlobalAllocCount() { 335 pthread_mutex_lock(&gParcelGlobalAllocSizeLock); 336 size_t count = gParcelGlobalAllocCount; 337 pthread_mutex_unlock(&gParcelGlobalAllocSizeLock); 338 return count; 339} 340 341const uint8_t* Parcel::data() const 342{ 343 return mData; 344} 345 346size_t Parcel::dataSize() const 347{ 348 return (mDataSize > mDataPos ? mDataSize : mDataPos); 349} 350 351size_t Parcel::dataAvail() const 352{ 353 // TODO: decide what to do about the possibility that this can 354 // report an available-data size that exceeds a Java int's max 355 // positive value, causing havoc. Fortunately this will only 356 // happen if someone constructs a Parcel containing more than two 357 // gigabytes of data, which on typical phone hardware is simply 358 // not possible. 359 return dataSize() - dataPosition(); 360} 361 362size_t Parcel::dataPosition() const 363{ 364 return mDataPos; 365} 366 367size_t Parcel::dataCapacity() const 368{ 369 return mDataCapacity; 370} 371 372status_t Parcel::setDataSize(size_t size) 373{ 374 if (size > INT32_MAX) { 375 // don't accept size_t values which may have come from an 376 // inadvertent conversion from a negative int. 377 return BAD_VALUE; 378 } 379 380 status_t err; 381 err = continueWrite(size); 382 if (err == NO_ERROR) { 383 mDataSize = size; 384 ALOGV("setDataSize Setting data size of %p to %zu", this, mDataSize); 385 } 386 return err; 387} 388 389void Parcel::setDataPosition(size_t pos) const 390{ 391 if (pos > INT32_MAX) { 392 // don't accept size_t values which may have come from an 393 // inadvertent conversion from a negative int. 394 abort(); 395 } 396 397 mDataPos = pos; 398 mNextObjectHint = 0; 399} 400 401status_t Parcel::setDataCapacity(size_t size) 402{ 403 if (size > INT32_MAX) { 404 // don't accept size_t values which may have come from an 405 // inadvertent conversion from a negative int. 406 return BAD_VALUE; 407 } 408 409 if (size > mDataCapacity) return continueWrite(size); 410 return NO_ERROR; 411} 412 413status_t Parcel::setData(const uint8_t* buffer, size_t len) 414{ 415 if (len > INT32_MAX) { 416 // don't accept size_t values which may have come from an 417 // inadvertent conversion from a negative int. 418 return BAD_VALUE; 419 } 420 421 status_t err = restartWrite(len); 422 if (err == NO_ERROR) { 423 memcpy(const_cast<uint8_t*>(data()), buffer, len); 424 mDataSize = len; 425 mFdsKnown = false; 426 } 427 return err; 428} 429 430status_t Parcel::appendFrom(const Parcel *parcel, size_t offset, size_t len) 431{ 432 const sp<ProcessState> proc(ProcessState::self()); 433 status_t err; 434 const uint8_t *data = parcel->mData; 435 const binder_size_t *objects = parcel->mObjects; 436 size_t size = parcel->mObjectsSize; 437 int startPos = mDataPos; 438 int firstIndex = -1, lastIndex = -2; 439 440 if (len == 0) { 441 return NO_ERROR; 442 } 443 444 if (len > INT32_MAX) { 445 // don't accept size_t values which may have come from an 446 // inadvertent conversion from a negative int. 447 return BAD_VALUE; 448 } 449 450 // range checks against the source parcel size 451 if ((offset > parcel->mDataSize) 452 || (len > parcel->mDataSize) 453 || (offset + len > parcel->mDataSize)) { 454 return BAD_VALUE; 455 } 456 457 // Count objects in range 458 for (int i = 0; i < (int) size; i++) { 459 size_t off = objects[i]; 460 if ((off >= offset) && (off + sizeof(flat_binder_object) <= offset + len)) { 461 if (firstIndex == -1) { 462 firstIndex = i; 463 } 464 lastIndex = i; 465 } 466 } 467 int numObjects = lastIndex - firstIndex + 1; 468 469 if ((mDataSize+len) > mDataCapacity) { 470 // grow data 471 err = growData(len); 472 if (err != NO_ERROR) { 473 return err; 474 } 475 } 476 477 // append data 478 memcpy(mData + mDataPos, data + offset, len); 479 mDataPos += len; 480 mDataSize += len; 481 482 err = NO_ERROR; 483 484 if (numObjects > 0) { 485 // grow objects 486 if (mObjectsCapacity < mObjectsSize + numObjects) { 487 size_t newSize = ((mObjectsSize + numObjects)*3)/2; 488 if (newSize < mObjectsSize) return NO_MEMORY; // overflow 489 binder_size_t *objects = 490 (binder_size_t*)realloc(mObjects, newSize*sizeof(binder_size_t)); 491 if (objects == (binder_size_t*)0) { 492 return NO_MEMORY; 493 } 494 mObjects = objects; 495 mObjectsCapacity = newSize; 496 } 497 498 // append and acquire objects 499 int idx = mObjectsSize; 500 for (int i = firstIndex; i <= lastIndex; i++) { 501 size_t off = objects[i] - offset + startPos; 502 mObjects[idx++] = off; 503 mObjectsSize++; 504 505 flat_binder_object* flat 506 = reinterpret_cast<flat_binder_object*>(mData + off); 507 acquire_object(proc, *flat, this); 508 509 if (flat->type == BINDER_TYPE_FD) { 510 // If this is a file descriptor, we need to dup it so the 511 // new Parcel now owns its own fd, and can declare that we 512 // officially know we have fds. 513 flat->handle = dup(flat->handle); 514 flat->cookie = 1; 515 mHasFds = mFdsKnown = true; 516 if (!mAllowFds) { 517 err = FDS_NOT_ALLOWED; 518 } 519 } 520 } 521 } 522 523 return err; 524} 525 526bool Parcel::allowFds() const 527{ 528 return mAllowFds; 529} 530 531bool Parcel::pushAllowFds(bool allowFds) 532{ 533 const bool origValue = mAllowFds; 534 if (!allowFds) { 535 mAllowFds = false; 536 } 537 return origValue; 538} 539 540void Parcel::restoreAllowFds(bool lastValue) 541{ 542 mAllowFds = lastValue; 543} 544 545bool Parcel::hasFileDescriptors() const 546{ 547 if (!mFdsKnown) { 548 scanForFds(); 549 } 550 return mHasFds; 551} 552 553// Write RPC headers. (previously just the interface token) 554status_t Parcel::writeInterfaceToken(const String16& interface) 555{ 556 writeInt32(IPCThreadState::self()->getStrictModePolicy() | 557 STRICT_MODE_PENALTY_GATHER); 558 // currently the interface identification token is just its name as a string 559 return writeString16(interface); 560} 561 562bool Parcel::checkInterface(IBinder* binder) const 563{ 564 return enforceInterface(binder->getInterfaceDescriptor()); 565} 566 567bool Parcel::enforceInterface(const String16& interface, 568 IPCThreadState* threadState) const 569{ 570 int32_t strictPolicy = readInt32(); 571 if (threadState == NULL) { 572 threadState = IPCThreadState::self(); 573 } 574 if ((threadState->getLastTransactionBinderFlags() & 575 IBinder::FLAG_ONEWAY) != 0) { 576 // For one-way calls, the callee is running entirely 577 // disconnected from the caller, so disable StrictMode entirely. 578 // Not only does disk/network usage not impact the caller, but 579 // there's no way to commuicate back any violations anyway. 580 threadState->setStrictModePolicy(0); 581 } else { 582 threadState->setStrictModePolicy(strictPolicy); 583 } 584 const String16 str(readString16()); 585 if (str == interface) { 586 return true; 587 } else { 588 ALOGW("**** enforceInterface() expected '%s' but read '%s'", 589 String8(interface).string(), String8(str).string()); 590 return false; 591 } 592} 593 594const binder_size_t* Parcel::objects() const 595{ 596 return mObjects; 597} 598 599size_t Parcel::objectsCount() const 600{ 601 return mObjectsSize; 602} 603 604status_t Parcel::errorCheck() const 605{ 606 return mError; 607} 608 609void Parcel::setError(status_t err) 610{ 611 mError = err; 612} 613 614status_t Parcel::finishWrite(size_t len) 615{ 616 if (len > INT32_MAX) { 617 // don't accept size_t values which may have come from an 618 // inadvertent conversion from a negative int. 619 return BAD_VALUE; 620 } 621 622 //printf("Finish write of %d\n", len); 623 mDataPos += len; 624 ALOGV("finishWrite Setting data pos of %p to %zu", this, mDataPos); 625 if (mDataPos > mDataSize) { 626 mDataSize = mDataPos; 627 ALOGV("finishWrite Setting data size of %p to %zu", this, mDataSize); 628 } 629 //printf("New pos=%d, size=%d\n", mDataPos, mDataSize); 630 return NO_ERROR; 631} 632 633status_t Parcel::writeUnpadded(const void* data, size_t len) 634{ 635 if (len > INT32_MAX) { 636 // don't accept size_t values which may have come from an 637 // inadvertent conversion from a negative int. 638 return BAD_VALUE; 639 } 640 641 size_t end = mDataPos + len; 642 if (end < mDataPos) { 643 // integer overflow 644 return BAD_VALUE; 645 } 646 647 if (end <= mDataCapacity) { 648restart_write: 649 memcpy(mData+mDataPos, data, len); 650 return finishWrite(len); 651 } 652 653 status_t err = growData(len); 654 if (err == NO_ERROR) goto restart_write; 655 return err; 656} 657 658status_t Parcel::write(const void* data, size_t len) 659{ 660 if (len > INT32_MAX) { 661 // don't accept size_t values which may have come from an 662 // inadvertent conversion from a negative int. 663 return BAD_VALUE; 664 } 665 666 void* const d = writeInplace(len); 667 if (d) { 668 memcpy(d, data, len); 669 return NO_ERROR; 670 } 671 return mError; 672} 673 674void* Parcel::writeInplace(size_t len) 675{ 676 if (len > INT32_MAX) { 677 // don't accept size_t values which may have come from an 678 // inadvertent conversion from a negative int. 679 return NULL; 680 } 681 682 const size_t padded = pad_size(len); 683 684 // sanity check for integer overflow 685 if (mDataPos+padded < mDataPos) { 686 return NULL; 687 } 688 689 if ((mDataPos+padded) <= mDataCapacity) { 690restart_write: 691 //printf("Writing %ld bytes, padded to %ld\n", len, padded); 692 uint8_t* const data = mData+mDataPos; 693 694 // Need to pad at end? 695 if (padded != len) { 696#if BYTE_ORDER == BIG_ENDIAN 697 static const uint32_t mask[4] = { 698 0x00000000, 0xffffff00, 0xffff0000, 0xff000000 699 }; 700#endif 701#if BYTE_ORDER == LITTLE_ENDIAN 702 static const uint32_t mask[4] = { 703 0x00000000, 0x00ffffff, 0x0000ffff, 0x000000ff 704 }; 705#endif 706 //printf("Applying pad mask: %p to %p\n", (void*)mask[padded-len], 707 // *reinterpret_cast<void**>(data+padded-4)); 708 *reinterpret_cast<uint32_t*>(data+padded-4) &= mask[padded-len]; 709 } 710 711 finishWrite(padded); 712 return data; 713 } 714 715 status_t err = growData(padded); 716 if (err == NO_ERROR) goto restart_write; 717 return NULL; 718} 719 720status_t Parcel::writeInt32(int32_t val) 721{ 722 return writeAligned(val); 723} 724 725status_t Parcel::writeUint32(uint32_t val) 726{ 727 return writeAligned(val); 728} 729 730status_t Parcel::writeInt32Array(size_t len, const int32_t *val) { 731 if (len > INT32_MAX) { 732 // don't accept size_t values which may have come from an 733 // inadvertent conversion from a negative int. 734 return BAD_VALUE; 735 } 736 737 if (!val) { 738 return writeInt32(-1); 739 } 740 status_t ret = writeInt32(static_cast<uint32_t>(len)); 741 if (ret == NO_ERROR) { 742 ret = write(val, len * sizeof(*val)); 743 } 744 return ret; 745} 746status_t Parcel::writeByteArray(size_t len, const uint8_t *val) { 747 if (len > INT32_MAX) { 748 // don't accept size_t values which may have come from an 749 // inadvertent conversion from a negative int. 750 return BAD_VALUE; 751 } 752 753 if (!val) { 754 return writeInt32(-1); 755 } 756 status_t ret = writeInt32(static_cast<uint32_t>(len)); 757 if (ret == NO_ERROR) { 758 ret = write(val, len * sizeof(*val)); 759 } 760 return ret; 761} 762 763status_t Parcel::writeInt64(int64_t val) 764{ 765 return writeAligned(val); 766} 767 768status_t Parcel::writeUint64(uint64_t val) 769{ 770 return writeAligned(val); 771} 772 773status_t Parcel::writePointer(uintptr_t val) 774{ 775 return writeAligned<binder_uintptr_t>(val); 776} 777 778status_t Parcel::writeFloat(float val) 779{ 780 return writeAligned(val); 781} 782 783#if defined(__mips__) && defined(__mips_hard_float) 784 785status_t Parcel::writeDouble(double val) 786{ 787 union { 788 double d; 789 unsigned long long ll; 790 } u; 791 u.d = val; 792 return writeAligned(u.ll); 793} 794 795#else 796 797status_t Parcel::writeDouble(double val) 798{ 799 return writeAligned(val); 800} 801 802#endif 803 804status_t Parcel::writeCString(const char* str) 805{ 806 return write(str, strlen(str)+1); 807} 808 809status_t Parcel::writeString8(const String8& str) 810{ 811 status_t err = writeInt32(str.bytes()); 812 // only write string if its length is more than zero characters, 813 // as readString8 will only read if the length field is non-zero. 814 // this is slightly different from how writeString16 works. 815 if (str.bytes() > 0 && err == NO_ERROR) { 816 err = write(str.string(), str.bytes()+1); 817 } 818 return err; 819} 820 821status_t Parcel::writeString16(const String16& str) 822{ 823 return writeString16(str.string(), str.size()); 824} 825 826status_t Parcel::writeString16(const char16_t* str, size_t len) 827{ 828 if (str == NULL) return writeInt32(-1); 829 830 status_t err = writeInt32(len); 831 if (err == NO_ERROR) { 832 len *= sizeof(char16_t); 833 uint8_t* data = (uint8_t*)writeInplace(len+sizeof(char16_t)); 834 if (data) { 835 memcpy(data, str, len); 836 *reinterpret_cast<char16_t*>(data+len) = 0; 837 return NO_ERROR; 838 } 839 err = mError; 840 } 841 return err; 842} 843 844status_t Parcel::writeStrongBinder(const sp<IBinder>& val) 845{ 846 return flatten_binder(ProcessState::self(), val, this); 847} 848 849status_t Parcel::writeWeakBinder(const wp<IBinder>& val) 850{ 851 return flatten_binder(ProcessState::self(), val, this); 852} 853 854status_t Parcel::writeNativeHandle(const native_handle* handle) 855{ 856 if (!handle || handle->version != sizeof(native_handle)) 857 return BAD_TYPE; 858 859 status_t err; 860 err = writeInt32(handle->numFds); 861 if (err != NO_ERROR) return err; 862 863 err = writeInt32(handle->numInts); 864 if (err != NO_ERROR) return err; 865 866 for (int i=0 ; err==NO_ERROR && i<handle->numFds ; i++) 867 err = writeDupFileDescriptor(handle->data[i]); 868 869 if (err != NO_ERROR) { 870 ALOGD("write native handle, write dup fd failed"); 871 return err; 872 } 873 err = write(handle->data + handle->numFds, sizeof(int)*handle->numInts); 874 return err; 875} 876 877status_t Parcel::writeFileDescriptor(int fd, bool takeOwnership) 878{ 879 flat_binder_object obj; 880 obj.type = BINDER_TYPE_FD; 881 obj.flags = 0x7f | FLAT_BINDER_FLAG_ACCEPTS_FDS; 882 obj.binder = 0; /* Don't pass uninitialized stack data to a remote process */ 883 obj.handle = fd; 884 obj.cookie = takeOwnership ? 1 : 0; 885 return writeObject(obj, true); 886} 887 888status_t Parcel::writeDupFileDescriptor(int fd) 889{ 890 int dupFd = dup(fd); 891 if (dupFd < 0) { 892 return -errno; 893 } 894 status_t err = writeFileDescriptor(dupFd, true /*takeOwnership*/); 895 if (err) { 896 close(dupFd); 897 } 898 return err; 899} 900 901status_t Parcel::writeBlob(size_t len, bool mutableCopy, WritableBlob* outBlob) 902{ 903 if (len > INT32_MAX) { 904 // don't accept size_t values which may have come from an 905 // inadvertent conversion from a negative int. 906 return BAD_VALUE; 907 } 908 909 status_t status; 910 if (!mAllowFds || len <= BLOB_INPLACE_LIMIT) { 911 ALOGV("writeBlob: write in place"); 912 status = writeInt32(BLOB_INPLACE); 913 if (status) return status; 914 915 void* ptr = writeInplace(len); 916 if (!ptr) return NO_MEMORY; 917 918 outBlob->init(-1, ptr, len, false); 919 return NO_ERROR; 920 } 921 922 ALOGV("writeBlob: write to ashmem"); 923 int fd = ashmem_create_region("Parcel Blob", len); 924 if (fd < 0) return NO_MEMORY; 925 926 mBlobAshmemSize += len; 927 928 int result = ashmem_set_prot_region(fd, PROT_READ | PROT_WRITE); 929 if (result < 0) { 930 status = result; 931 } else { 932 void* ptr = ::mmap(NULL, len, PROT_READ | PROT_WRITE, MAP_SHARED, fd, 0); 933 if (ptr == MAP_FAILED) { 934 status = -errno; 935 } else { 936 if (!mutableCopy) { 937 result = ashmem_set_prot_region(fd, PROT_READ); 938 } 939 if (result < 0) { 940 status = result; 941 } else { 942 status = writeInt32(mutableCopy ? BLOB_ASHMEM_MUTABLE : BLOB_ASHMEM_IMMUTABLE); 943 if (!status) { 944 status = writeFileDescriptor(fd, true /*takeOwnership*/); 945 if (!status) { 946 outBlob->init(fd, ptr, len, mutableCopy); 947 return NO_ERROR; 948 } 949 } 950 } 951 } 952 ::munmap(ptr, len); 953 } 954 ::close(fd); 955 return status; 956} 957 958status_t Parcel::writeDupImmutableBlobFileDescriptor(int fd) 959{ 960 // Must match up with what's done in writeBlob. 961 if (!mAllowFds) return FDS_NOT_ALLOWED; 962 status_t status = writeInt32(BLOB_ASHMEM_IMMUTABLE); 963 if (status) return status; 964 return writeDupFileDescriptor(fd); 965} 966 967status_t Parcel::write(const FlattenableHelperInterface& val) 968{ 969 status_t err; 970 971 // size if needed 972 const size_t len = val.getFlattenedSize(); 973 const size_t fd_count = val.getFdCount(); 974 975 if ((len > INT32_MAX) || (fd_count > INT32_MAX)) { 976 // don't accept size_t values which may have come from an 977 // inadvertent conversion from a negative int. 978 return BAD_VALUE; 979 } 980 981 err = this->writeInt32(len); 982 if (err) return err; 983 984 err = this->writeInt32(fd_count); 985 if (err) return err; 986 987 // payload 988 void* const buf = this->writeInplace(pad_size(len)); 989 if (buf == NULL) 990 return BAD_VALUE; 991 992 int* fds = NULL; 993 if (fd_count) { 994 fds = new int[fd_count]; 995 } 996 997 err = val.flatten(buf, len, fds, fd_count); 998 for (size_t i=0 ; i<fd_count && err==NO_ERROR ; i++) { 999 err = this->writeDupFileDescriptor( fds[i] ); 1000 } 1001 1002 if (fd_count) { 1003 delete [] fds; 1004 } 1005 1006 return err; 1007} 1008 1009status_t Parcel::writeObject(const flat_binder_object& val, bool nullMetaData) 1010{ 1011 const bool enoughData = (mDataPos+sizeof(val)) <= mDataCapacity; 1012 const bool enoughObjects = mObjectsSize < mObjectsCapacity; 1013 if (enoughData && enoughObjects) { 1014restart_write: 1015 *reinterpret_cast<flat_binder_object*>(mData+mDataPos) = val; 1016 1017 // remember if it's a file descriptor 1018 if (val.type == BINDER_TYPE_FD) { 1019 if (!mAllowFds) { 1020 // fail before modifying our object index 1021 return FDS_NOT_ALLOWED; 1022 } 1023 mHasFds = mFdsKnown = true; 1024 } 1025 1026 // Need to write meta-data? 1027 if (nullMetaData || val.binder != 0) { 1028 mObjects[mObjectsSize] = mDataPos; 1029 acquire_object(ProcessState::self(), val, this); 1030 mObjectsSize++; 1031 } 1032 1033 return finishWrite(sizeof(flat_binder_object)); 1034 } 1035 1036 if (!enoughData) { 1037 const status_t err = growData(sizeof(val)); 1038 if (err != NO_ERROR) return err; 1039 } 1040 if (!enoughObjects) { 1041 size_t newSize = ((mObjectsSize+2)*3)/2; 1042 if (newSize < mObjectsSize) return NO_MEMORY; // overflow 1043 binder_size_t* objects = (binder_size_t*)realloc(mObjects, newSize*sizeof(binder_size_t)); 1044 if (objects == NULL) return NO_MEMORY; 1045 mObjects = objects; 1046 mObjectsCapacity = newSize; 1047 } 1048 1049 goto restart_write; 1050} 1051 1052status_t Parcel::writeNoException() 1053{ 1054 return writeInt32(0); 1055} 1056 1057void Parcel::remove(size_t /*start*/, size_t /*amt*/) 1058{ 1059 LOG_ALWAYS_FATAL("Parcel::remove() not yet implemented!"); 1060} 1061 1062status_t Parcel::read(void* outData, size_t len) const 1063{ 1064 if (len > INT32_MAX) { 1065 // don't accept size_t values which may have come from an 1066 // inadvertent conversion from a negative int. 1067 return BAD_VALUE; 1068 } 1069 1070 if ((mDataPos+pad_size(len)) >= mDataPos && (mDataPos+pad_size(len)) <= mDataSize 1071 && len <= pad_size(len)) { 1072 memcpy(outData, mData+mDataPos, len); 1073 mDataPos += pad_size(len); 1074 ALOGV("read Setting data pos of %p to %zu", this, mDataPos); 1075 return NO_ERROR; 1076 } 1077 return NOT_ENOUGH_DATA; 1078} 1079 1080const void* Parcel::readInplace(size_t len) const 1081{ 1082 if (len > INT32_MAX) { 1083 // don't accept size_t values which may have come from an 1084 // inadvertent conversion from a negative int. 1085 return NULL; 1086 } 1087 1088 if ((mDataPos+pad_size(len)) >= mDataPos && (mDataPos+pad_size(len)) <= mDataSize 1089 && len <= pad_size(len)) { 1090 const void* data = mData+mDataPos; 1091 mDataPos += pad_size(len); 1092 ALOGV("readInplace Setting data pos of %p to %zu", this, mDataPos); 1093 return data; 1094 } 1095 return NULL; 1096} 1097 1098template<class T> 1099status_t Parcel::readAligned(T *pArg) const { 1100 COMPILE_TIME_ASSERT_FUNCTION_SCOPE(PAD_SIZE_UNSAFE(sizeof(T)) == sizeof(T)); 1101 1102 if ((mDataPos+sizeof(T)) <= mDataSize) { 1103 const void* data = mData+mDataPos; 1104 mDataPos += sizeof(T); 1105 *pArg = *reinterpret_cast<const T*>(data); 1106 return NO_ERROR; 1107 } else { 1108 return NOT_ENOUGH_DATA; 1109 } 1110} 1111 1112template<class T> 1113T Parcel::readAligned() const { 1114 T result; 1115 if (readAligned(&result) != NO_ERROR) { 1116 result = 0; 1117 } 1118 1119 return result; 1120} 1121 1122template<class T> 1123status_t Parcel::writeAligned(T val) { 1124 COMPILE_TIME_ASSERT_FUNCTION_SCOPE(PAD_SIZE_UNSAFE(sizeof(T)) == sizeof(T)); 1125 1126 if ((mDataPos+sizeof(val)) <= mDataCapacity) { 1127restart_write: 1128 *reinterpret_cast<T*>(mData+mDataPos) = val; 1129 return finishWrite(sizeof(val)); 1130 } 1131 1132 status_t err = growData(sizeof(val)); 1133 if (err == NO_ERROR) goto restart_write; 1134 return err; 1135} 1136 1137status_t Parcel::readInt32(int32_t *pArg) const 1138{ 1139 return readAligned(pArg); 1140} 1141 1142int32_t Parcel::readInt32() const 1143{ 1144 return readAligned<int32_t>(); 1145} 1146 1147status_t Parcel::readUint32(uint32_t *pArg) const 1148{ 1149 return readAligned(pArg); 1150} 1151 1152uint32_t Parcel::readUint32() const 1153{ 1154 return readAligned<uint32_t>(); 1155} 1156 1157status_t Parcel::readInt64(int64_t *pArg) const 1158{ 1159 return readAligned(pArg); 1160} 1161 1162 1163int64_t Parcel::readInt64() const 1164{ 1165 return readAligned<int64_t>(); 1166} 1167 1168status_t Parcel::readUint64(uint64_t *pArg) const 1169{ 1170 return readAligned(pArg); 1171} 1172 1173uint64_t Parcel::readUint64() const 1174{ 1175 return readAligned<uint64_t>(); 1176} 1177 1178status_t Parcel::readPointer(uintptr_t *pArg) const 1179{ 1180 status_t ret; 1181 binder_uintptr_t ptr; 1182 ret = readAligned(&ptr); 1183 if (!ret) 1184 *pArg = ptr; 1185 return ret; 1186} 1187 1188uintptr_t Parcel::readPointer() const 1189{ 1190 return readAligned<binder_uintptr_t>(); 1191} 1192 1193 1194status_t Parcel::readFloat(float *pArg) const 1195{ 1196 return readAligned(pArg); 1197} 1198 1199 1200float Parcel::readFloat() const 1201{ 1202 return readAligned<float>(); 1203} 1204 1205#if defined(__mips__) && defined(__mips_hard_float) 1206 1207status_t Parcel::readDouble(double *pArg) const 1208{ 1209 union { 1210 double d; 1211 unsigned long long ll; 1212 } u; 1213 u.d = 0; 1214 status_t status; 1215 status = readAligned(&u.ll); 1216 *pArg = u.d; 1217 return status; 1218} 1219 1220double Parcel::readDouble() const 1221{ 1222 union { 1223 double d; 1224 unsigned long long ll; 1225 } u; 1226 u.ll = readAligned<unsigned long long>(); 1227 return u.d; 1228} 1229 1230#else 1231 1232status_t Parcel::readDouble(double *pArg) const 1233{ 1234 return readAligned(pArg); 1235} 1236 1237double Parcel::readDouble() const 1238{ 1239 return readAligned<double>(); 1240} 1241 1242#endif 1243 1244status_t Parcel::readIntPtr(intptr_t *pArg) const 1245{ 1246 return readAligned(pArg); 1247} 1248 1249 1250intptr_t Parcel::readIntPtr() const 1251{ 1252 return readAligned<intptr_t>(); 1253} 1254 1255 1256const char* Parcel::readCString() const 1257{ 1258 const size_t avail = mDataSize-mDataPos; 1259 if (avail > 0) { 1260 const char* str = reinterpret_cast<const char*>(mData+mDataPos); 1261 // is the string's trailing NUL within the parcel's valid bounds? 1262 const char* eos = reinterpret_cast<const char*>(memchr(str, 0, avail)); 1263 if (eos) { 1264 const size_t len = eos - str; 1265 mDataPos += pad_size(len+1); 1266 ALOGV("readCString Setting data pos of %p to %zu", this, mDataPos); 1267 return str; 1268 } 1269 } 1270 return NULL; 1271} 1272 1273String8 Parcel::readString8() const 1274{ 1275 int32_t size = readInt32(); 1276 // watch for potential int overflow adding 1 for trailing NUL 1277 if (size > 0 && size < INT32_MAX) { 1278 const char* str = (const char*)readInplace(size+1); 1279 if (str) return String8(str, size); 1280 } 1281 return String8(); 1282} 1283 1284String16 Parcel::readString16() const 1285{ 1286 size_t len; 1287 const char16_t* str = readString16Inplace(&len); 1288 if (str) return String16(str, len); 1289 ALOGE("Reading a NULL string not supported here."); 1290 return String16(); 1291} 1292 1293const char16_t* Parcel::readString16Inplace(size_t* outLen) const 1294{ 1295 int32_t size = readInt32(); 1296 // watch for potential int overflow from size+1 1297 if (size >= 0 && size < INT32_MAX) { 1298 *outLen = size; 1299 const char16_t* str = (const char16_t*)readInplace((size+1)*sizeof(char16_t)); 1300 if (str != NULL) { 1301 return str; 1302 } 1303 } 1304 *outLen = 0; 1305 return NULL; 1306} 1307 1308sp<IBinder> Parcel::readStrongBinder() const 1309{ 1310 sp<IBinder> val; 1311 unflatten_binder(ProcessState::self(), *this, &val); 1312 return val; 1313} 1314 1315wp<IBinder> Parcel::readWeakBinder() const 1316{ 1317 wp<IBinder> val; 1318 unflatten_binder(ProcessState::self(), *this, &val); 1319 return val; 1320} 1321 1322int32_t Parcel::readExceptionCode() const 1323{ 1324 int32_t exception_code = readAligned<int32_t>(); 1325 if (exception_code == EX_HAS_REPLY_HEADER) { 1326 int32_t header_start = dataPosition(); 1327 int32_t header_size = readAligned<int32_t>(); 1328 // Skip over fat responses headers. Not used (or propagated) in 1329 // native code 1330 setDataPosition(header_start + header_size); 1331 // And fat response headers are currently only used when there are no 1332 // exceptions, so return no error: 1333 return 0; 1334 } 1335 return exception_code; 1336} 1337 1338native_handle* Parcel::readNativeHandle() const 1339{ 1340 int numFds, numInts; 1341 status_t err; 1342 err = readInt32(&numFds); 1343 if (err != NO_ERROR) return 0; 1344 err = readInt32(&numInts); 1345 if (err != NO_ERROR) return 0; 1346 1347 native_handle* h = native_handle_create(numFds, numInts); 1348 if (!h) { 1349 return 0; 1350 } 1351 1352 for (int i=0 ; err==NO_ERROR && i<numFds ; i++) { 1353 h->data[i] = dup(readFileDescriptor()); 1354 if (h->data[i] < 0) { 1355 for (int j = 0; j < i; j++) { 1356 close(h->data[j]); 1357 } 1358 native_handle_delete(h); 1359 return 0; 1360 } 1361 } 1362 err = read(h->data + numFds, sizeof(int)*numInts); 1363 if (err != NO_ERROR) { 1364 native_handle_close(h); 1365 native_handle_delete(h); 1366 h = 0; 1367 } 1368 return h; 1369} 1370 1371 1372int Parcel::readFileDescriptor() const 1373{ 1374 const flat_binder_object* flat = readObject(true); 1375 if (flat) { 1376 switch (flat->type) { 1377 case BINDER_TYPE_FD: 1378 //ALOGI("Returning file descriptor %ld from parcel %p", flat->handle, this); 1379 return flat->handle; 1380 } 1381 } 1382 return BAD_TYPE; 1383} 1384 1385status_t Parcel::readBlob(size_t len, ReadableBlob* outBlob) const 1386{ 1387 int32_t blobType; 1388 status_t status = readInt32(&blobType); 1389 if (status) return status; 1390 1391 if (blobType == BLOB_INPLACE) { 1392 ALOGV("readBlob: read in place"); 1393 const void* ptr = readInplace(len); 1394 if (!ptr) return BAD_VALUE; 1395 1396 outBlob->init(-1, const_cast<void*>(ptr), len, false); 1397 return NO_ERROR; 1398 } 1399 1400 ALOGV("readBlob: read from ashmem"); 1401 bool isMutable = (blobType == BLOB_ASHMEM_MUTABLE); 1402 int fd = readFileDescriptor(); 1403 if (fd == int(BAD_TYPE)) return BAD_VALUE; 1404 1405 void* ptr = ::mmap(NULL, len, isMutable ? PROT_READ | PROT_WRITE : PROT_READ, 1406 MAP_SHARED, fd, 0); 1407 if (ptr == MAP_FAILED) return NO_MEMORY; 1408 1409 outBlob->init(fd, ptr, len, isMutable); 1410 return NO_ERROR; 1411} 1412 1413status_t Parcel::read(FlattenableHelperInterface& val) const 1414{ 1415 // size 1416 const size_t len = this->readInt32(); 1417 const size_t fd_count = this->readInt32(); 1418 1419 if (len > INT32_MAX) { 1420 // don't accept size_t values which may have come from an 1421 // inadvertent conversion from a negative int. 1422 return BAD_VALUE; 1423 } 1424 1425 // payload 1426 void const* const buf = this->readInplace(pad_size(len)); 1427 if (buf == NULL) 1428 return BAD_VALUE; 1429 1430 int* fds = NULL; 1431 if (fd_count) { 1432 fds = new int[fd_count]; 1433 } 1434 1435 status_t err = NO_ERROR; 1436 for (size_t i=0 ; i<fd_count && err==NO_ERROR ; i++) { 1437 fds[i] = dup(this->readFileDescriptor()); 1438 if (fds[i] < 0) { 1439 err = BAD_VALUE; 1440 ALOGE("dup() failed in Parcel::read, i is %zu, fds[i] is %d, fd_count is %zu, error: %s", 1441 i, fds[i], fd_count, strerror(errno)); 1442 } 1443 } 1444 1445 if (err == NO_ERROR) { 1446 err = val.unflatten(buf, len, fds, fd_count); 1447 } 1448 1449 if (fd_count) { 1450 delete [] fds; 1451 } 1452 1453 return err; 1454} 1455const flat_binder_object* Parcel::readObject(bool nullMetaData) const 1456{ 1457 const size_t DPOS = mDataPos; 1458 if ((DPOS+sizeof(flat_binder_object)) <= mDataSize) { 1459 const flat_binder_object* obj 1460 = reinterpret_cast<const flat_binder_object*>(mData+DPOS); 1461 mDataPos = DPOS + sizeof(flat_binder_object); 1462 if (!nullMetaData && (obj->cookie == 0 && obj->binder == 0)) { 1463 // When transferring a NULL object, we don't write it into 1464 // the object list, so we don't want to check for it when 1465 // reading. 1466 ALOGV("readObject Setting data pos of %p to %zu", this, mDataPos); 1467 return obj; 1468 } 1469 1470 // Ensure that this object is valid... 1471 binder_size_t* const OBJS = mObjects; 1472 const size_t N = mObjectsSize; 1473 size_t opos = mNextObjectHint; 1474 1475 if (N > 0) { 1476 ALOGV("Parcel %p looking for obj at %zu, hint=%zu", 1477 this, DPOS, opos); 1478 1479 // Start at the current hint position, looking for an object at 1480 // the current data position. 1481 if (opos < N) { 1482 while (opos < (N-1) && OBJS[opos] < DPOS) { 1483 opos++; 1484 } 1485 } else { 1486 opos = N-1; 1487 } 1488 if (OBJS[opos] == DPOS) { 1489 // Found it! 1490 ALOGV("Parcel %p found obj %zu at index %zu with forward search", 1491 this, DPOS, opos); 1492 mNextObjectHint = opos+1; 1493 ALOGV("readObject Setting data pos of %p to %zu", this, mDataPos); 1494 return obj; 1495 } 1496 1497 // Look backwards for it... 1498 while (opos > 0 && OBJS[opos] > DPOS) { 1499 opos--; 1500 } 1501 if (OBJS[opos] == DPOS) { 1502 // Found it! 1503 ALOGV("Parcel %p found obj %zu at index %zu with backward search", 1504 this, DPOS, opos); 1505 mNextObjectHint = opos+1; 1506 ALOGV("readObject Setting data pos of %p to %zu", this, mDataPos); 1507 return obj; 1508 } 1509 } 1510 ALOGW("Attempt to read object from Parcel %p at offset %zu that is not in the object list", 1511 this, DPOS); 1512 } 1513 return NULL; 1514} 1515 1516void Parcel::closeFileDescriptors() 1517{ 1518 size_t i = mObjectsSize; 1519 if (i > 0) { 1520 //ALOGI("Closing file descriptors for %zu objects...", i); 1521 } 1522 while (i > 0) { 1523 i--; 1524 const flat_binder_object* flat 1525 = reinterpret_cast<flat_binder_object*>(mData+mObjects[i]); 1526 if (flat->type == BINDER_TYPE_FD) { 1527 //ALOGI("Closing fd: %ld", flat->handle); 1528 close(flat->handle); 1529 } 1530 } 1531} 1532 1533uintptr_t Parcel::ipcData() const 1534{ 1535 return reinterpret_cast<uintptr_t>(mData); 1536} 1537 1538size_t Parcel::ipcDataSize() const 1539{ 1540 return (mDataSize > mDataPos ? mDataSize : mDataPos); 1541} 1542 1543uintptr_t Parcel::ipcObjects() const 1544{ 1545 return reinterpret_cast<uintptr_t>(mObjects); 1546} 1547 1548size_t Parcel::ipcObjectsCount() const 1549{ 1550 return mObjectsSize; 1551} 1552 1553void Parcel::ipcSetDataReference(const uint8_t* data, size_t dataSize, 1554 const binder_size_t* objects, size_t objectsCount, release_func relFunc, void* relCookie) 1555{ 1556 binder_size_t minOffset = 0; 1557 freeDataNoInit(); 1558 mError = NO_ERROR; 1559 mData = const_cast<uint8_t*>(data); 1560 mDataSize = mDataCapacity = dataSize; 1561 //ALOGI("setDataReference Setting data size of %p to %lu (pid=%d)", this, mDataSize, getpid()); 1562 mDataPos = 0; 1563 ALOGV("setDataReference Setting data pos of %p to %zu", this, mDataPos); 1564 mObjects = const_cast<binder_size_t*>(objects); 1565 mObjectsSize = mObjectsCapacity = objectsCount; 1566 mNextObjectHint = 0; 1567 mOwner = relFunc; 1568 mOwnerCookie = relCookie; 1569 for (size_t i = 0; i < mObjectsSize; i++) { 1570 binder_size_t offset = mObjects[i]; 1571 if (offset < minOffset) { 1572 ALOGE("%s: bad object offset %" PRIu64 " < %" PRIu64 "\n", 1573 __func__, (uint64_t)offset, (uint64_t)minOffset); 1574 mObjectsSize = 0; 1575 break; 1576 } 1577 minOffset = offset + sizeof(flat_binder_object); 1578 } 1579 scanForFds(); 1580} 1581 1582void Parcel::print(TextOutput& to, uint32_t /*flags*/) const 1583{ 1584 to << "Parcel("; 1585 1586 if (errorCheck() != NO_ERROR) { 1587 const status_t err = errorCheck(); 1588 to << "Error: " << (void*)(intptr_t)err << " \"" << strerror(-err) << "\""; 1589 } else if (dataSize() > 0) { 1590 const uint8_t* DATA = data(); 1591 to << indent << HexDump(DATA, dataSize()) << dedent; 1592 const binder_size_t* OBJS = objects(); 1593 const size_t N = objectsCount(); 1594 for (size_t i=0; i<N; i++) { 1595 const flat_binder_object* flat 1596 = reinterpret_cast<const flat_binder_object*>(DATA+OBJS[i]); 1597 to << endl << "Object #" << i << " @ " << (void*)OBJS[i] << ": " 1598 << TypeCode(flat->type & 0x7f7f7f00) 1599 << " = " << flat->binder; 1600 } 1601 } else { 1602 to << "NULL"; 1603 } 1604 1605 to << ")"; 1606} 1607 1608void Parcel::releaseObjects() 1609{ 1610 const sp<ProcessState> proc(ProcessState::self()); 1611 size_t i = mObjectsSize; 1612 uint8_t* const data = mData; 1613 binder_size_t* const objects = mObjects; 1614 while (i > 0) { 1615 i--; 1616 const flat_binder_object* flat 1617 = reinterpret_cast<flat_binder_object*>(data+objects[i]); 1618 release_object(proc, *flat, this); 1619 } 1620} 1621 1622void Parcel::acquireObjects() 1623{ 1624 const sp<ProcessState> proc(ProcessState::self()); 1625 size_t i = mObjectsSize; 1626 uint8_t* const data = mData; 1627 binder_size_t* const objects = mObjects; 1628 while (i > 0) { 1629 i--; 1630 const flat_binder_object* flat 1631 = reinterpret_cast<flat_binder_object*>(data+objects[i]); 1632 acquire_object(proc, *flat, this); 1633 } 1634} 1635 1636void Parcel::freeData() 1637{ 1638 freeDataNoInit(); 1639 initState(); 1640} 1641 1642void Parcel::freeDataNoInit() 1643{ 1644 if (mOwner) { 1645 LOG_ALLOC("Parcel %p: freeing other owner data", this); 1646 //ALOGI("Freeing data ref of %p (pid=%d)", this, getpid()); 1647 mOwner(this, mData, mDataSize, mObjects, mObjectsSize, mOwnerCookie); 1648 } else { 1649 LOG_ALLOC("Parcel %p: freeing allocated data", this); 1650 releaseObjects(); 1651 if (mData) { 1652 LOG_ALLOC("Parcel %p: freeing with %zu capacity", this, mDataCapacity); 1653 pthread_mutex_lock(&gParcelGlobalAllocSizeLock); 1654 gParcelGlobalAllocSize -= mDataCapacity; 1655 gParcelGlobalAllocCount--; 1656 pthread_mutex_unlock(&gParcelGlobalAllocSizeLock); 1657 free(mData); 1658 } 1659 if (mObjects) free(mObjects); 1660 } 1661} 1662 1663status_t Parcel::growData(size_t len) 1664{ 1665 if (len > INT32_MAX) { 1666 // don't accept size_t values which may have come from an 1667 // inadvertent conversion from a negative int. 1668 return BAD_VALUE; 1669 } 1670 1671 size_t newSize = ((mDataSize+len)*3)/2; 1672 return (newSize <= mDataSize) 1673 ? (status_t) NO_MEMORY 1674 : continueWrite(newSize); 1675} 1676 1677status_t Parcel::restartWrite(size_t desired) 1678{ 1679 if (desired > INT32_MAX) { 1680 // don't accept size_t values which may have come from an 1681 // inadvertent conversion from a negative int. 1682 return BAD_VALUE; 1683 } 1684 1685 if (mOwner) { 1686 freeData(); 1687 return continueWrite(desired); 1688 } 1689 1690 uint8_t* data = (uint8_t*)realloc(mData, desired); 1691 if (!data && desired > mDataCapacity) { 1692 mError = NO_MEMORY; 1693 return NO_MEMORY; 1694 } 1695 1696 releaseObjects(); 1697 1698 if (data) { 1699 LOG_ALLOC("Parcel %p: restart from %zu to %zu capacity", this, mDataCapacity, desired); 1700 pthread_mutex_lock(&gParcelGlobalAllocSizeLock); 1701 gParcelGlobalAllocSize += desired; 1702 gParcelGlobalAllocSize -= mDataCapacity; 1703 pthread_mutex_unlock(&gParcelGlobalAllocSizeLock); 1704 mData = data; 1705 mDataCapacity = desired; 1706 } 1707 1708 mDataSize = mDataPos = 0; 1709 ALOGV("restartWrite Setting data size of %p to %zu", this, mDataSize); 1710 ALOGV("restartWrite Setting data pos of %p to %zu", this, mDataPos); 1711 1712 free(mObjects); 1713 mObjects = NULL; 1714 mObjectsSize = mObjectsCapacity = 0; 1715 mNextObjectHint = 0; 1716 mHasFds = false; 1717 mFdsKnown = true; 1718 mAllowFds = true; 1719 1720 return NO_ERROR; 1721} 1722 1723status_t Parcel::continueWrite(size_t desired) 1724{ 1725 if (desired > INT32_MAX) { 1726 // don't accept size_t values which may have come from an 1727 // inadvertent conversion from a negative int. 1728 return BAD_VALUE; 1729 } 1730 1731 // If shrinking, first adjust for any objects that appear 1732 // after the new data size. 1733 size_t objectsSize = mObjectsSize; 1734 if (desired < mDataSize) { 1735 if (desired == 0) { 1736 objectsSize = 0; 1737 } else { 1738 while (objectsSize > 0) { 1739 if (mObjects[objectsSize-1] < desired) 1740 break; 1741 objectsSize--; 1742 } 1743 } 1744 } 1745 1746 if (mOwner) { 1747 // If the size is going to zero, just release the owner's data. 1748 if (desired == 0) { 1749 freeData(); 1750 return NO_ERROR; 1751 } 1752 1753 // If there is a different owner, we need to take 1754 // posession. 1755 uint8_t* data = (uint8_t*)malloc(desired); 1756 if (!data) { 1757 mError = NO_MEMORY; 1758 return NO_MEMORY; 1759 } 1760 binder_size_t* objects = NULL; 1761 1762 if (objectsSize) { 1763 objects = (binder_size_t*)calloc(objectsSize, sizeof(binder_size_t)); 1764 if (!objects) { 1765 free(data); 1766 1767 mError = NO_MEMORY; 1768 return NO_MEMORY; 1769 } 1770 1771 // Little hack to only acquire references on objects 1772 // we will be keeping. 1773 size_t oldObjectsSize = mObjectsSize; 1774 mObjectsSize = objectsSize; 1775 acquireObjects(); 1776 mObjectsSize = oldObjectsSize; 1777 } 1778 1779 if (mData) { 1780 memcpy(data, mData, mDataSize < desired ? mDataSize : desired); 1781 } 1782 if (objects && mObjects) { 1783 memcpy(objects, mObjects, objectsSize*sizeof(binder_size_t)); 1784 } 1785 //ALOGI("Freeing data ref of %p (pid=%d)", this, getpid()); 1786 mOwner(this, mData, mDataSize, mObjects, mObjectsSize, mOwnerCookie); 1787 mOwner = NULL; 1788 1789 LOG_ALLOC("Parcel %p: taking ownership of %zu capacity", this, desired); 1790 pthread_mutex_lock(&gParcelGlobalAllocSizeLock); 1791 gParcelGlobalAllocSize += desired; 1792 gParcelGlobalAllocCount++; 1793 pthread_mutex_unlock(&gParcelGlobalAllocSizeLock); 1794 1795 mData = data; 1796 mObjects = objects; 1797 mDataSize = (mDataSize < desired) ? mDataSize : desired; 1798 ALOGV("continueWrite Setting data size of %p to %zu", this, mDataSize); 1799 mDataCapacity = desired; 1800 mObjectsSize = mObjectsCapacity = objectsSize; 1801 mNextObjectHint = 0; 1802 1803 } else if (mData) { 1804 if (objectsSize < mObjectsSize) { 1805 // Need to release refs on any objects we are dropping. 1806 const sp<ProcessState> proc(ProcessState::self()); 1807 for (size_t i=objectsSize; i<mObjectsSize; i++) { 1808 const flat_binder_object* flat 1809 = reinterpret_cast<flat_binder_object*>(mData+mObjects[i]); 1810 if (flat->type == BINDER_TYPE_FD) { 1811 // will need to rescan because we may have lopped off the only FDs 1812 mFdsKnown = false; 1813 } 1814 release_object(proc, *flat, this); 1815 } 1816 binder_size_t* objects = 1817 (binder_size_t*)realloc(mObjects, objectsSize*sizeof(binder_size_t)); 1818 if (objects) { 1819 mObjects = objects; 1820 } 1821 mObjectsSize = objectsSize; 1822 mNextObjectHint = 0; 1823 } 1824 1825 // We own the data, so we can just do a realloc(). 1826 if (desired > mDataCapacity) { 1827 uint8_t* data = (uint8_t*)realloc(mData, desired); 1828 if (data) { 1829 LOG_ALLOC("Parcel %p: continue from %zu to %zu capacity", this, mDataCapacity, 1830 desired); 1831 pthread_mutex_lock(&gParcelGlobalAllocSizeLock); 1832 gParcelGlobalAllocSize += desired; 1833 gParcelGlobalAllocSize -= mDataCapacity; 1834 pthread_mutex_unlock(&gParcelGlobalAllocSizeLock); 1835 mData = data; 1836 mDataCapacity = desired; 1837 } else if (desired > mDataCapacity) { 1838 mError = NO_MEMORY; 1839 return NO_MEMORY; 1840 } 1841 } else { 1842 if (mDataSize > desired) { 1843 mDataSize = desired; 1844 ALOGV("continueWrite Setting data size of %p to %zu", this, mDataSize); 1845 } 1846 if (mDataPos > desired) { 1847 mDataPos = desired; 1848 ALOGV("continueWrite Setting data pos of %p to %zu", this, mDataPos); 1849 } 1850 } 1851 1852 } else { 1853 // This is the first data. Easy! 1854 uint8_t* data = (uint8_t*)malloc(desired); 1855 if (!data) { 1856 mError = NO_MEMORY; 1857 return NO_MEMORY; 1858 } 1859 1860 if(!(mDataCapacity == 0 && mObjects == NULL 1861 && mObjectsCapacity == 0)) { 1862 ALOGE("continueWrite: %zu/%p/%zu/%zu", mDataCapacity, mObjects, mObjectsCapacity, desired); 1863 } 1864 1865 LOG_ALLOC("Parcel %p: allocating with %zu capacity", this, desired); 1866 pthread_mutex_lock(&gParcelGlobalAllocSizeLock); 1867 gParcelGlobalAllocSize += desired; 1868 gParcelGlobalAllocCount++; 1869 pthread_mutex_unlock(&gParcelGlobalAllocSizeLock); 1870 1871 mData = data; 1872 mDataSize = mDataPos = 0; 1873 ALOGV("continueWrite Setting data size of %p to %zu", this, mDataSize); 1874 ALOGV("continueWrite Setting data pos of %p to %zu", this, mDataPos); 1875 mDataCapacity = desired; 1876 } 1877 1878 return NO_ERROR; 1879} 1880 1881void Parcel::initState() 1882{ 1883 LOG_ALLOC("Parcel %p: initState", this); 1884 mError = NO_ERROR; 1885 mData = 0; 1886 mDataSize = 0; 1887 mDataCapacity = 0; 1888 mDataPos = 0; 1889 ALOGV("initState Setting data size of %p to %zu", this, mDataSize); 1890 ALOGV("initState Setting data pos of %p to %zu", this, mDataPos); 1891 mObjects = NULL; 1892 mObjectsSize = 0; 1893 mObjectsCapacity = 0; 1894 mNextObjectHint = 0; 1895 mHasFds = false; 1896 mFdsKnown = true; 1897 mAllowFds = true; 1898 mOwner = NULL; 1899 mBlobAshmemSize = 0; 1900} 1901 1902void Parcel::scanForFds() const 1903{ 1904 bool hasFds = false; 1905 for (size_t i=0; i<mObjectsSize; i++) { 1906 const flat_binder_object* flat 1907 = reinterpret_cast<const flat_binder_object*>(mData + mObjects[i]); 1908 if (flat->type == BINDER_TYPE_FD) { 1909 hasFds = true; 1910 break; 1911 } 1912 } 1913 mHasFds = hasFds; 1914 mFdsKnown = true; 1915} 1916 1917size_t Parcel::getBlobAshmemSize() const 1918{ 1919 return mBlobAshmemSize; 1920} 1921 1922// --- Parcel::Blob --- 1923 1924Parcel::Blob::Blob() : 1925 mFd(-1), mData(NULL), mSize(0), mMutable(false) { 1926} 1927 1928Parcel::Blob::~Blob() { 1929 release(); 1930} 1931 1932void Parcel::Blob::release() { 1933 if (mFd != -1 && mData) { 1934 ::munmap(mData, mSize); 1935 } 1936 clear(); 1937} 1938 1939void Parcel::Blob::init(int fd, void* data, size_t size, bool isMutable) { 1940 mFd = fd; 1941 mData = data; 1942 mSize = size; 1943 mMutable = isMutable; 1944} 1945 1946void Parcel::Blob::clear() { 1947 mFd = -1; 1948 mData = NULL; 1949 mSize = 0; 1950 mMutable = false; 1951} 1952 1953}; // namespace android 1954