/external/tensorflow/tensorflow/contrib/boosted_trees/lib/learner/batch/ |
H A D | base_split_handler.py | 71 def update_stats(self, stamp_token, example_partition_ids, gradients, 77 stamp_token: An int32 scalar tensor containing the current stamp token. 96 def update_stats_sync(self, stamp_token, example_partition_ids, gradients, 102 stamp_token: An int32 scalar tensor containing the current stamp token. 120 handler_reads, stamp_token, None) 122 stamp_token, example_partition_ids, gradients, hessians, 127 }, stamp_token, None) 131 def make_splits(self, stamp_token, next_stamp_token, class_id): 137 stamp_token: An int32 scalar tensor containing the current stamp token.
|
H A D | ordinal_split_handler.py | 198 def update_stats(self, stamp_token, example_partition_ids, gradients, 204 stamp_token: An int32 scalar tensor containing the current stamp token. 230 stamp_token=stamp_token, 237 def make_splits(self, stamp_token, next_stamp_token, class_id): 241 self._quantile_accumulator.get_buckets(stamp_token)) 246 stamp_token=stamp_token, next_stamp_token=next_stamp_token) 255 self._stats_accumulator.flush(stamp_token, next_stamp_token)) 338 def update_stats(self, stamp_token, example_partition_id [all...] |
H A D | categorical_split_handler.py | 84 def update_stats(self, stamp_token, example_partition_ids, gradients, 90 stamp_token: An int32 scalar tensor containing the current stamp token. 170 def make_splits(self, stamp_token, next_stamp_token, class_id): 175 self._stats_accumulator.flush(stamp_token, next_stamp_token))
|
/external/tensorflow/tensorflow/contrib/boosted_trees/python/ops/ |
H A D | quantile_ops.py | 86 stamp_token, state, are_buckets_ready, buckets = ( 95 specs = [make_save_spec(stamp_token, "_stamp")] 116 stamp_token = restored_tensors[:1] 123 stamp_token=stamp_token, 128 def get_buckets(self, stamp_token): 133 stamp_token=stamp_token)) 160 def add_summary(self, stamp_token, column, example_weights): 165 stamp_token [all...] |
H A D | model_ops.py | 50 stamp_token, ensemble_config = tree_ensemble_serialize(tree_ensemble_handle) 56 saver.BaseSaverBuilder.SaveSpec(stamp_token, slice_spec, 80 stamp_token=restored_tensors[0], 84 def tree_ensemble_variable(stamp_token, 91 stamp_token: The initial stamp token value for the ensemble resource. 104 resource_handle, stamp_token, tree_ensemble_config)
|
H A D | stats_accumulator_ops.py | 44 stamp_token, 52 stamp_token: An int64, initial value to use for the stamp token. 73 self._resource_handle, stamp_token) 83 self._resource_handle, stamp_token, gradient_shape.as_list(), 92 (stamp_token, num_updates, partition_ids, feature_ids, gradients, 95 saver.BaseSaverBuilder.SaveSpec(stamp_token, slice_spec, 114 def add(self, stamp_token, partition_ids, feature_ids, gradients, hessians): 120 [self._resource_handle], stamp_token, [partition_ids], [feature_ids], 124 [self._resource_handle], stamp_token, [partition_ids], [feature_ids], 156 def deserialize(self, stamp_token, num_update [all...] |
/external/tensorflow/tensorflow/contrib/boosted_trees/python/kernel_tests/ |
H A D | stats_accumulator_ops_test.py | 34 stamp_token=0, 39 stamp_token=0, 48 stamp_token=0, next_stamp_token=1) 62 stamp_token=0, 67 stamp_token=0, 76 stamp_token=0, next_stamp_token=1) 91 stamp_token=0, 96 stamp_token=0, 102 stamp_token=-1, 110 stamp_token [all...] |
H A D | model_ops_test.py | 112 stamp_token=3, 131 stamp_token = model_ops.tree_ensemble_stamp_token(tree_ensemble_handle) 132 self.assertEqual(stamp_token.eval(), 3) 157 stamp_token=7, 160 stamp_token, serialized_config = model_ops.tree_ensemble_serialize( 163 self.assertEqual(stamp_token.eval(), 7) 169 stamp_token=9, 226 stamp_token=3, 241 stamp_token=3, 270 stamp_token [all...] |
H A D | training_ops_test.py | 132 stamp_token=0, 155 stamp_token=0, 168 training_ops.tree_ensemble_stats(tree_ensemble_handle, stamp_token=1)) 206 stamp_token=1, 217 training_ops.tree_ensemble_stats(tree_ensemble_handle, stamp_token=2)) 253 stamp_token=2, 264 training_ops.tree_ensemble_stats(tree_ensemble_handle, stamp_token=3)) 283 stamp_token=0, 316 stamp_token=0, 334 training_ops.tree_ensemble_stats(tree_ensemble_handle, stamp_token [all...] |
H A D | quantile_ops_test.py | 118 stamp_token=0, 131 sess.run(accumulator.flush(stamp_token=0, next_stamp_token=1)) 132 are_ready_flush, buckets = (accumulator.get_buckets(stamp_token=1)) 153 stamp_token=0, 178 sess.run(accumulator.flush(stamp_token=0, next_stamp_token=1)) 179 are_ready_flush, buckets = (accumulator.get_buckets(stamp_token=1)) 200 stamp_token=0, 210 sess.run(accumulator.flush(stamp_token=0, next_stamp_token=1)) 211 are_ready_flush, buckets = (accumulator.get_buckets(stamp_token=1)) 275 stamp_token [all...] |
H A D | prediction_ops_test.py | 174 stamp_token=0, 201 stamp_token=0, 231 stamp_token=0, 275 stamp_token=0, 350 stamp_token=0, 412 stamp_token=0, 459 stamp_token=0, 509 stamp_token=0, 555 stamp_token=0, 606 stamp_token [all...] |
/external/tensorflow/tensorflow/contrib/boosted_trees/python/training/functions/ |
H A D | gbdt_batch_test.py | 149 stamp_token=0, tree_ensemble_config="", name="tree_ensemble") 200 stamp_token, serialized = model_ops.tree_ensemble_serialize( 206 self.assertEquals(stamp_token.eval(), 1) 214 stamp_token, serialized = model_ops.tree_ensemble_serialize( 220 self.assertEquals(stamp_token.eval(), 2) 252 stamp_token=0, tree_ensemble_config="", name="tree_ensemble") 304 stamp_token, serialized = model_ops.tree_ensemble_serialize( 310 self.assertEquals(stamp_token.eval(), 1) 318 stamp_token, serialized = model_ops.tree_ensemble_serialize( 324 self.assertEquals(stamp_token [all...] |
H A D | gbdt_batch.py | 434 local_ensemble_handle, stamp_token=-1, tree_ensemble_config="") 450 stamp_token=ensemble_stamp, 631 stamp_token=0, 638 stamp_token=0, 958 stamp_token=ensemble_stamp, 973 stamp_token=ensemble_stamp, 987 stamp_token=ensemble_stamp, 1013 self._ensemble_handle, stamp_token=next_ensemble_stamp)
|
/external/tensorflow/tensorflow/contrib/boosted_trees/resources/ |
H A D | quantile_stream_resource.h | 36 int64 stamp_token) 43 set_stamp(stamp_token); 34 QuantileStreamResource(const float epsilon, const int32 num_quantiles, const int64 max_elements, bool generate_quantiles, int64 stamp_token) argument
|
H A D | decision_tree_ensemble_resource.h | 49 bool InitFromSerialized(const string& serialized, const int64 stamp_token) { argument 52 set_stamp(stamp_token);
|
/external/tensorflow/tensorflow/contrib/boosted_trees/kernels/ |
H A D | model_ops.cc | 40 OP_REQUIRES_OK(context, context->input("stamp_token", &stamp_token_t)); 41 int64 stamp_token = stamp_token_t->scalar<int64>()(); variable 49 stamp_token)) { 123 OP_REQUIRES_OK(context, context->input("stamp_token", &stamp_token_t)); 124 int64 stamp_token = stamp_token_t->scalar<int64>()(); variable 135 tree_ensemble_config_t->scalar<string>()(), stamp_token),
|
H A D | quantile_ops.cc | 46 const char* const kStampTokenName = "stamp_token"; 281 int64 stamp_token = stamp_token_t->scalar<int64>()(); variable 288 [&context, &resource_handle_list, &summary_list, stamp_token]( 303 if (!streams_resource->is_stamp_valid(stamp_token)) { 306 << " Passed stamp token: " << stamp_token << " " 329 streams_resource->stream(stamp_token)->PushSummary(entries); 470 int64 stamp_token = streams_resource->stamp(); variable 479 const QuantileStream& stream = *streams_resource->stream(stamp_token); 481 are_buckets_ready ? streams_resource->boundaries(stamp_token) 503 stamp_token_t->scalar<int64>()() = stamp_token; 569 int64 stamp_token = stamp_token_t->scalar<int64>()(); variable 605 int64 stamp_token = stamp_token_t->scalar<int64>()(); variable 651 int64 stamp_token = stamp_token_t->scalar<int64>()(); variable 694 int64 stamp_token = stamp_token_t->scalar<int64>()(); variable [all...] |
H A D | stats_accumulator_ops.cc | 38 const char* const kStampTokenName = "stamp_token"; 362 OP_REQUIRES_OK(context, context->input("stamp_token", &stamp_token_t)); 389 OP_REQUIRES_OK(context, context->input("stamp_token", &stamp_token_t)); 439 int64 stamp_token = stamp_token_t->scalar<int64>()(); variable 448 stamp_token](int64 start, int64 end) { 461 if (!accumulator_resource->is_stamp_valid(stamp_token)) { 463 << "Passed stamp token: " << stamp_token << " " 503 int64 stamp_token = stamp_token_t->scalar<int64>()(); variable 512 stamp_token](int64 start, int64 end) { 525 if (!accumulator_resource->is_stamp_valid(stamp_token)) { 558 int64 stamp_token = stamp_token_t->scalar<int64>()(); variable 600 int64 stamp_token = stamp_token_t->scalar<int64>()(); variable 641 int64 stamp_token = stamp_token_t->scalar<int64>()(); variable 670 int64 stamp_token = stamp_token_t->scalar<int64>()(); variable [all...] |
H A D | training_ops.cc | 197 OP_REQUIRES_OK(context, context->input("stamp_token", &stamp_token_t)); 198 int64 stamp_token = stamp_token_t->scalar<int64>()(); variable 202 CHECK(ensemble_resource->is_stamp_valid(stamp_token)); 209 CHECK(stamp_token != next_stamp_token); 323 OP_REQUIRES_OK(context, context->input("stamp_token", &stamp_token_t)); 324 int64 stamp_token = stamp_token_t->scalar<int64>()(); variable 328 CHECK(ensemble_resource->is_stamp_valid(stamp_token)); 335 CHECK(stamp_token != next_stamp_token); 750 OP_REQUIRES_OK(context, context->input("stamp_token", &stamp_token_t)); 751 int64 stamp_token variable [all...] |
/external/tensorflow/tensorflow/contrib/boosted_trees/estimator_batch/ |
H A D | model.py | 82 stamp_token=0,
|
H A D | dnn_tree_combined_estimator.py | 187 stamp_token=0,
|