@@ -555,10 +555,12 @@ void FleetWrapper::PushSparseFromTensorAsync(
555
555
// in
556
556
// ctr_accessor.h
557
557
push_values.back ()[0 ] = 2 ; // TODO(zhaocaibei123): slot
558
- push_values.back ()[1 ] =
559
- (i >= show_size ? 1 : static_cast <float >(show_tensor[i]));
560
- push_values.back ()[2 ] =
561
- (i >= clk_size ? 0 : static_cast <float >(clk_tensor[i]));
558
+ push_values.back ()[1 ] = (static_cast <int >(i) >= show_size
559
+ ? 1
560
+ : static_cast <float >(show_tensor[i]));
561
+ push_values.back ()[2 ] = (static_cast <int >(i) >= clk_size
562
+ ? 0
563
+ : static_cast <float >(clk_tensor[i]));
562
564
float * data = push_values.back ().data () + 3 ;
563
565
memcpy (data, g + output_len, sizeof (float ) * fea_dim);
564
566
}
@@ -582,10 +584,12 @@ void FleetWrapper::PushSparseFromTensorAsync(
582
584
// slot show clk grad... consistent with CtrCommonPushValue defined in
583
585
// ctr_accessor.h
584
586
push_values.back ()[0 ] = 2 ; // TODO(zhaocaibei123): slot
585
- push_values.back ()[1 ] =
586
- (i >= show_size ? 1 : static_cast <float >(show_tensor[i]));
587
- push_values.back ()[2 ] =
588
- (i >= clk_size ? 0 : static_cast <float >(clk_tensor[i]));
587
+ push_values.back ()[1 ] = (static_cast <int >(i) >= show_size
588
+ ? 1
589
+ : static_cast <float >(show_tensor[i]));
590
+ push_values.back ()[2 ] = (static_cast <int >(i) >= clk_size
591
+ ? 0
592
+ : static_cast <float >(clk_tensor[i]));
589
593
float * data = push_values.back ().data () + 3 ;
590
594
memcpy (data, g + output_len, sizeof (float ) * fea_dim);
591
595
}
0 commit comments