/external/webrtc/tools/python_charts/data/ |
H A D | vp8_hw.py | 40 frame_data = [{'frame_number': 0, 'encoding_successful': True , 'decoding_successful': True , 'encode_time': 94676, 'decode_time': 37942, 'encode_return_code': 0, 'decode_return_code': 0, 'bit_rate': 1098, 'encoded_frame_length': 4579, 'frame_type': 'Other', 'packets_dropped': 0, 'total_packets': 4, 'ssim': 0.910364, 'psnr': 35.067258}, variable
|
H A D | vp8_sw.py | 40 frame_data = [{'frame_number': 0, 'encoding_successful': True , 'decoding_successful': True , 'encode_time': 12427, 'decode_time': 4403, 'encode_return_code': 0, 'decode_return_code': 0, 'bit_rate': 2270, 'encoded_frame_length': 9459, 'frame_type': 'Other', 'packets_dropped': 0, 'total_packets': 7, 'ssim': 0.947050, 'psnr': 38.332820}, variable
|
/external/webrtc/webrtc/modules/video_coding/ |
H A D | decoding_state_unittest.cc | 41 FrameData frame_data; local 42 frame_data.rtt_ms = 0; 43 frame_data.rolling_average_packets_per_frame = -1; 44 EXPECT_LE(0, frame.InsertPacket(packet, 0, kNoErrors, frame_data)); 49 EXPECT_LE(0, frame_key.InsertPacket(packet, 0, kNoErrors, frame_data)); 57 EXPECT_LE(0, frame.InsertPacket(packet, 0, kNoErrors, frame_data)); 62 EXPECT_LE(0, frame.InsertPacket(packet, 0, kNoErrors, frame_data)); 69 EXPECT_LE(0, frame.InsertPacket(packet, 0, kNoErrors, frame_data)); 73 EXPECT_LE(0, frame.InsertPacket(packet, 0, kNoErrors, frame_data)); 76 EXPECT_LE(0, frame.InsertPacket(packet, 0, kNoErrors, frame_data)); 170 FrameData frame_data; local 220 FrameData frame_data; local 374 FrameData frame_data; local 403 FrameData frame_data; local 427 FrameData frame_data; local 465 FrameData frame_data; local 508 FrameData frame_data; local 563 FrameData frame_data; local [all...] |
H A D | session_info_unittest.cc | 32 frame_data.rtt_ms = 0; 33 frame_data.rolling_average_packets_per_frame = -1; 61 FrameData frame_data; member in class:webrtc::TestSessionInfo 157 kNoErrors, frame_data))); 166 kNoErrors, frame_data))); 179 0, session_.InsertPacket(packet_, frame_buffer_, kNoErrors, frame_data)); 190 kNoErrors, frame_data))); 198 packet_, frame_buffer_, kNoErrors, frame_data))); 206 kNoErrors, frame_data))); 222 packet_, frame_buffer_, kWithErrors, frame_data))); [all...] |
H A D | frame_buffer.cc | 87 const FrameData& frame_data) { 134 _sessionInfo.InsertPacket(packet, _buffer, decode_error_mode, frame_data); 83 InsertPacket( const VCMPacket& packet, int64_t timeInMs, VCMDecodeErrorMode decode_error_mode, const FrameData& frame_data) argument
|
H A D | session_info.cc | 261 void VCMSessionInfo::UpdateDecodableSession(const FrameData& frame_data) { argument 273 if (frame_data.rtt_ms < kRttThreshold || frame_type_ == kVideoFrameKey || 276 frame_data.rolling_average_packets_per_frame && 278 frame_data.rolling_average_packets_per_frame)) 473 const FrameData& frame_data) { 551 UpdateDecodableSession(frame_data); 470 InsertPacket(const VCMPacket& packet, uint8_t* frame_buffer, VCMDecodeErrorMode decode_error_mode, const FrameData& frame_data) argument
|
H A D | jitter_buffer.cc | 737 FrameData frame_data; local 738 frame_data.rtt_ms = rtt_ms_; 739 frame_data.rolling_average_packets_per_frame = average_packets_per_frame_; 741 frame->InsertPacket(packet, now_ms, decode_error_mode_, frame_data);
|
/external/autotest/client/site_tests/graphics_WebGLManyPlanetsDeep/ |
H A D | graphics_WebGLManyPlanetsDeep.py | 21 frame_data = {} variable in class:graphics_WebGLManyPlanetsDeep 57 frame_data = tab.EvaluateJavaScript( 59 for datum in frame_data: 60 if not datum or datum['seq'] in self.frame_data: 62 self.frame_data[datum['seq']] = { 77 for v in self.frame_data.itervalues()]) 107 with open('frame_data', 'w') as f: 111 for k in sorted(self.frame_data.keys()): 112 d = self.frame_data[k]
|
/external/libvpx/libvpx/test/ |
H A D | svc_test.cc | 255 uint8_t *frame_data = static_cast<uint8_t *>(inputs[i].buf); local 256 uint8_t *frame_start = frame_data; 259 if (frame_data[0] & 0x02) { 263 frame_data += frame_sizes[frame]; 271 frame_data += frame_sizes[frame]; 283 memmove(frame_data + 1, frame_start + inputs[i].sz - index_sz + 1, 286 frame_data[0] = marker; 287 frame_data += (mag * (frame + 1) + 1); 289 *frame_data++ = marker; 290 inputs[i].sz = frame_data [all...] |
/external/tensorflow/tensorflow/examples/android/jni/object_tracking/ |
H A D | object_tracker_jni.cc | 66 jfloat x2, jfloat y2, jbyteArray frame_data); 172 jfloat x2, jfloat y2, jbyteArray frame_data) { 181 jbyte* pixels = env->GetByteArrayElements(frame_data, &iCopied); 187 env->ReleaseByteArrayElements(frame_data, pixels, JNI_ABORT); 170 registerNewObjectWithAppearanceNative( JNIEnv* env, jobject thiz, jstring object_id, jfloat x1, jfloat y1, jfloat x2, jfloat y2, jbyteArray frame_data) argument
|
/external/webp/src/mux/ |
H A D | muxread.c | 442 const WebPData* frame_data; local 446 frame_data = &wpi->header_->data_; 447 if (frame_data->size < kChunks[IDX_ANMF].size) return WEBP_MUX_BAD_DATA; 449 frame->x_offset = 2 * GetLE24(frame_data->bytes + 0); 450 frame->y_offset = 2 * GetLE24(frame_data->bytes + 3); 452 const uint8_t bits = frame_data->bytes[15]; 453 frame->duration = GetLE24(frame_data->bytes + 12);
|
/external/webrtc/webrtc/modules/audio_coding/codecs/opus/ |
H A D | opus_interface.c | 454 const unsigned char *frame_data[48]; local 490 frame_data, frame_sizes, NULL) < 0) { 499 if (frame_data[0][0] & (0x80 >> ((n + 1) * (frames + 1) - 1)))
|
/external/libxaac/decoder/ |
H A D | ixheaacd_esbr_envcal.c | 71 VOID ixheaacd_sbr_env_calc(ia_sbr_frame_info_data_struct *frame_data, argument 86 WORD32 upsamp_4_flag = frame_data->pstr_sbr_header->is_usf_4; 98 FLOAT32 *prev_env_noise_level = frame_data->prev_noise_level; 107 FLOAT32 *sfb_nrg = frame_data->flt_env_sf_arr; 108 FLOAT32 *noise_floor = frame_data->flt_noise_floor; 109 ia_frame_info_struct *p_frame_info = &frame_data->str_frame_info_details; 111 ia_frame_info_struct *pvc_frame_info = &frame_data->str_pvc_frame_info; 112 WORD32 smoothing_length = frame_data->pstr_sbr_header->smoothing_mode ? 0 : 4; 113 WORD32 int_mode = frame_data->pstr_sbr_header->interpol_freq; 114 WORD32 limiter_band = frame_data [all...] |
/external/tensorflow/tensorflow/contrib/pi_examples/camera/ |
H A D | camera.cc | 194 uint8_t** frame_data, int* frame_data_size, 222 *frame_data = static_cast<uint8_t*>(buffers[buf->index].start); 480 uint8_t* frame_data; local 484 camera_handle, buffers, &frame_data, &frame_data_size, &buf); 492 TensorFromFrame(frame_data, video_width, video_height, 3, input_height, 193 CaptureNextFrame(int camera_handle, CameraBuffer* buffers, uint8_t** frame_data, int* frame_data_size, v4l2_buffer* buf) argument
|
/external/google-breakpad/src/common/windows/ |
H A D | pdb_source_line_writer.cc | 515 CComPtr<IDiaFrameData> frame_data; local 517 while (SUCCEEDED(frame_data_enum->Next(1, &frame_data, &count)) && 520 if (FAILED(frame_data->get_type(&type))) 524 if (FAILED(frame_data->get_relativeVirtualAddress(&rva))) 528 if (FAILED(frame_data->get_lengthBlock(&code_size))) 532 if (FAILED(frame_data->get_lengthProlog(&prolog_size))) 539 if (FAILED(frame_data->get_lengthParams(¶meter_size))) 543 if (FAILED(frame_data->get_lengthSavedRegisters(&saved_register_size))) 547 if (FAILED(frame_data->get_lengthLocals(&local_size))) 552 if (FAILED(frame_data [all...] |