Lines Matching refs:get

143     [m_objcObserver.get() disconnect];
151 [[NSNotificationCenter defaultCenter] removeObserver:m_objcObserver.get()];
156 [m_avAsset.get() cancelLoading];
161 [m_avPlayerItem.get() removeObserver:m_objcObserver.get() forKeyPath:keyName];
167 [m_avPlayer.get() removeTimeObserver:m_timeObserver];
168 [m_avPlayer.get() removeObserver:m_objcObserver.get() forKeyPath:@"rate"];
191 m_imageGenerator = [AVAssetImageGenerator assetImageGeneratorWithAsset:m_avAsset.get()];
193 [m_imageGenerator.get() setApertureMode:AVAssetImageGeneratorApertureModeCleanAperture];
194 [m_imageGenerator.get() setAppliesPreferredTrackTransform:YES];
196 LOG(Media, "MediaPlayerPrivateAVFoundationObjC::createImageGenerator(%p) - returning %p", this, m_imageGenerator.get());
204 LOG(Media, "MediaPlayerPrivateAVFoundationObjC::destroyContextVideoRenderer(%p) - destroying %p", this, m_imageGenerator.get());
216 [m_videoLayer.get() setPlayer:m_avPlayer.get()];
217 LOG(Media, "MediaPlayerPrivateAVFoundationObjC::createVideoLayer(%p) - returning %p", this, m_videoLayer.get());
226 LOG(Media, "MediaPlayerPrivateAVFoundationObjC::destroyVideoLayer(%p) - destroying", this, m_videoLayer.get());
228 [m_videoLayer.get() setPlayer:nil];
235 return (m_videoLayer && [m_videoLayer.get() isReadyForDisplay]);
276 [m_avPlayer.get() addObserver:m_objcObserver.get() forKeyPath:@"rate" options:nil context:(void *)MediaPlayerAVFoundationObservationContextPlayer];
281 WebCoreAVFMovieObserver *observer = m_objcObserver.get();
282 m_timeObserver = [m_avPlayer.get() addPeriodicTimeObserverForInterval:CMTimeMakeWithSeconds(veryLongInterval, 10) queue:nil usingBlock:^(CMTime time){
289 m_avPlayerItem.adoptNS([[AVPlayerItem alloc] initWithAsset:m_avAsset.get()]);
291 [[NSNotificationCenter defaultCenter] addObserver:m_objcObserver.get()selector:@selector(didEnd:) name:AVPlayerItemDidPlayToEndTimeNotification object:m_avPlayerItem.get()];
294 [m_avPlayerItem.get() addObserver:m_objcObserver.get() forKeyPath:keyName options:nil context:(void *)MediaPlayerAVFoundationObservationContextPlayerItem];
296 [m_avPlayer.get() replaceCurrentItemWithPlayerItem:m_avPlayerItem.get()];
306 [m_avAsset.get() loadValuesAsynchronouslyForKeys:[NSArray arrayWithObject:@"playable"] completionHandler:^{
307 [m_objcObserver.get() playableKnown];
314 [m_avAsset.get() loadValuesAsynchronouslyForKeys:[assetMetadataKeyNames() retain] completionHandler:^{
315 [m_objcObserver.get() metadataLoaded];
324 AVPlayerItemStatus status = [m_avPlayerItem.get() status];
329 if ([m_avPlayerItem.get() isPlaybackLikelyToKeepUp])
344 pm.media.avfMediaPlayer = m_avPlayer.get();
351 return m_videoLayer.get();
361 [m_avPlayer.get() setRate:requestedRate()];
372 [m_avPlayer.get() setRate:nil];
382 CMTime cmDuration = [m_avPlayerItem.get() duration];
400 CMTime itemTime = [m_avPlayerItem.get() currentTime];
412 WebCoreAVFMovieObserver *observer = m_objcObserver.get();
413 [m_avPlayerItem.get() seekToTime:CMTimeMakeWithSeconds(time, 600) toleranceBefore:kCMTimeZero toleranceAfter:kCMTimeZero completionHandler:^(BOOL finished) {
425 [m_avPlayer.get() setVolume:volume];
434 [m_avPlayer.get() setClosedCaptionDisplayEnabled:closedCaptionsVisible];
440 [m_avPlayer.get() setRate:requestedRate()];
449 return [m_avPlayer.get() rate];
459 NSArray *loadedRanges = [m_avPlayerItem.get() loadedTimeRanges];
473 NSArray *seekableRanges = [m_avPlayerItem.get() seekableTimeRanges];
492 NSArray *loadedRanges = [m_avPlayerItem.get() loadedTimeRanges];
516 NSArray *tracks = [m_avAsset.get() tracks];
534 AVKeyValueStatus keyStatus = [m_avAsset.get() statusOfValueForKey:keyName error:nil];
544 if ([[m_avAsset.get() valueForKey:@"playable"] boolValue])
573 CGContextDrawImage(context->platformContext(), CGRectMake(0, 0, paintRect.width(), paintRect.height()), image.get());
610 [m_imageGenerator.get() setMaximumSize:CGSize(rect.size())];
611 CGImageRef image = [m_imageGenerator.get() copyCGImageAtTime:CMTimeMakeWithSeconds(time, 600) actualTime:nil error:nil];
656 // This is called whenever the tracks collection changes so cache hasVideo and hasAudio since we get
658 setHasVideo([[m_avAsset.get() tracksWithMediaCharacteristic:AVMediaCharacteristicVisual] count]);
659 setHasAudio([[m_avAsset.get() tracksWithMediaCharacteristic:AVMediaCharacteristicAudible] count]);
660 setHasClosedCaptions([[m_avAsset.get() tracksWithMediaType:AVMediaTypeClosedCaption] count]);
667 NSArray *tracks = [m_avAsset.get() tracks];
672 setNaturalSize(IntSize([m_avPlayerItem.get() presentationSize]));
689 CGSize naturalSize = CGSizeApplyAffineTransform(trackUnionRect.size, [m_avAsset.get() preferredTransform]);