@@ -3303,9 +3303,9 @@ struct _SetRangeImpl
33033303{
33043304 struct _Data
33053305 {
3306- _DifferenceType __pos{}; // Offset in output range w/o limitation to output data size
3307- _DifferenceType __len{}; // Length in temporary buffer w/o limitation to output data size
3308- _DifferenceType __buf_pos{}; // Position in temporary buffer w/o limitation to output data size
3306+ _DifferenceType __pos{}; // Offset in output range w/o limitation to output data size
3307+ _DifferenceType __len{}; // Length in temporary buffer w/o limitation to output data size
3308+ _DifferenceType __buf_pos{}; // Position in temporary buffer w/o limitation to output data size
33093309
33103310 inline bool
33113311 empty () const
@@ -3356,8 +3356,7 @@ struct _SetRangeCombiner
33563356 }
33573357 else
33583358 {
3359- return {__a.__data [0 ].combine_with (__b.__data [0 ]),
3360- __a.__data [1 ].combine_with (__b.__data [1 ])};
3359+ return {__a.__data [0 ].combine_with (__b.__data [0 ]), __a.__data [1 ].combine_with (__b.__data [1 ])};
33613360 }
33623361 }
33633362};
@@ -3366,16 +3365,12 @@ struct _SetRangeCombiner
33663365// that output buffer is enough to keep all output data and all input data will be processed
33673366struct __set_op_unbounded_offsets_eval
33683367{
3369- template <class _IsVector , class _ExecutionPolicy ,
3370- typename _DifferenceType1, typename _DifferenceType2, typename _DifferenceTypeOut,
3371- class _SizeFunction , class _MaskSizeFunction , typename _MaskIterator>
3368+ template <class _IsVector , class _ExecutionPolicy , typename _DifferenceType1, typename _DifferenceType2,
3369+ typename _DifferenceTypeOut, class _SizeFunction , class _MaskSizeFunction , typename _MaskIterator>
33723370 std::pair<_DifferenceType1, _DifferenceType2>
3373- operator ()(__parallel_tag<_IsVector>, _ExecutionPolicy&&,
3374- _DifferenceType1 __n1, _DifferenceType2 __n2, [[maybe_unused]] _DifferenceTypeOut __n_out,
3375- [[maybe_unused]] _SizeFunction __size_func,
3376- _MaskSizeFunction,
3377- _MaskIterator,
3378- _DifferenceTypeOut) const
3371+ operator ()(__parallel_tag<_IsVector>, _ExecutionPolicy&&, _DifferenceType1 __n1, _DifferenceType2 __n2,
3372+ [[maybe_unused]] _DifferenceTypeOut __n_out, [[maybe_unused]] _SizeFunction __size_func,
3373+ _MaskSizeFunction, _MaskIterator, _DifferenceTypeOut) const
33793374 {
33803375 assert (__size_func (__n1, __n2) <= __n_out);
33813376
@@ -3417,9 +3412,7 @@ struct __mask_buffers<true>
34173412template <>
34183413struct __mask_buffers <false >
34193414{
3420- __mask_buffers (std::size_t )
3421- {
3422- }
3415+ __mask_buffers (std::size_t ) {}
34233416
34243417 std::nullptr_t
34253418 get_buf_mask_rng_data (std::size_t = 0 ) const
@@ -3461,14 +3454,12 @@ struct _ScanPred
34613454 }
34623455
34633456 protected:
3464-
34653457 template <typename _Data, typename ItFrom, typename ItTo>
34663458 void
34673459 __move_data_no_bounds (_Data __data, ItFrom __from, ItTo __to) const
34683460 {
3469- __brick_move_destroy<__parallel_tag<_IsVector>>{}(__from + __data.__buf_pos ,
3470- __from + __data.__buf_pos + __data.__len ,
3471- __to + __data.__pos , _IsVector{});
3461+ __brick_move_destroy<__parallel_tag<_IsVector>>{}(
3462+ __from + __data.__buf_pos , __from + __data.__buf_pos + __data.__len , __to + __data.__pos , _IsVector{});
34723463 }
34733464
34743465 template <typename _Data>
@@ -3512,10 +3503,12 @@ struct _ScanPred
35123503
35133504 // Evaluate pointers to current data chunk in temporary buffer
35143505 const auto __buf_raw_data_from = __advance_clamped (__buf_raw_data_begin, __data.__buf_pos , __buf_raw_data_end);
3515- const auto __buf_raw_data_to = __advance_clamped (__buf_raw_data_begin, __data.__buf_pos + std::min (__result_remaining, __data.__len ), __buf_raw_data_end);
3506+ const auto __buf_raw_data_to = __advance_clamped (
3507+ __buf_raw_data_begin, __data.__buf_pos + std::min (__result_remaining, __data.__len ), __buf_raw_data_end);
35163508
35173509 // Copy results data into results range to have final output
3518- __brick_move_destroy<__parallel_tag<_IsVector>>{}(__buf_raw_data_from, __buf_raw_data_to, __result_from, _IsVector{});
3510+ __brick_move_destroy<__parallel_tag<_IsVector>>{}(__buf_raw_data_from, __buf_raw_data_to, __result_from,
3511+ _IsVector{});
35193512 }
35203513
35213514 // Move it1 forward by n, but not beyond it2
@@ -3535,7 +3528,7 @@ struct _ParallelSetOpStrictScanPred
35353528{
35363529 _RandomAccessIterator1 __first1, __last1;
35373530 _RandomAccessIterator2 __first2, __last2;
3538- _SizeFunction __size_func; // Processing set data size function
3531+ _SizeFunction __size_func; // Processing set data size function
35393532 _SetUnionOp __set_union_op;
35403533
35413534 _Compare __comp;
@@ -3583,8 +3576,9 @@ struct _ParallelSetOpStrictScanPred
35833576 }
35843577 else
35853578 {
3586- return _SetRange{typename _SetRange::_Data{0 , 0 , __size_func ((__b - __first1), (__bb - __first2))},
3587- typename _SetRange::_Data{0 , 0 , __mask_size_func ((__b - __first1), (__bb - __first2))}};
3579+ return _SetRange{
3580+ typename _SetRange::_Data{0 , 0 , __size_func ((__b - __first1), (__bb - __first2))},
3581+ typename _SetRange::_Data{0 , 0 , __mask_size_func ((__b - __first1), (__bb - __first2))}};
35883582 }
35893583 }
35903584
@@ -3674,9 +3668,12 @@ __parallel_set_op(__parallel_tag<_IsVector> __tag, _ExecutionPolicy&& __exec, _R
36743668 // Scan predicate
36753669 _ScanPred<__Bounded, _IsVector, decltype (__buf_raw_data_begin), decltype (__buf_mask_rng_raw_data_begin),
36763670 _OutputIterator>
3677- __scan_pred{__buf_raw_data_begin, __buf_raw_data_end,
3678- __buf_mask_rng_raw_data_begin, __buf_mask_rng_res_raw_data_begin,
3679- __result1, __result2};
3671+ __scan_pred{__buf_raw_data_begin,
3672+ __buf_raw_data_end,
3673+ __buf_mask_rng_raw_data_begin,
3674+ __buf_mask_rng_res_raw_data_begin,
3675+ __result1,
3676+ __result2};
36803677
36813678 _ParallelSetOpStrictScanPred<__Bounded, _SetRange, _RandomAccessIterator1, _RandomAccessIterator2,
36823679 _OutputIterator, _SizeFunction, _SetUnionOp, _Compare, _Proj1, _Proj2, _T>
@@ -3764,7 +3761,8 @@ __parallel_set_union_op(__parallel_tag<_IsVector> __tag, _ExecutionPolicy&& __ex
37643761 auto __last1_tmp = !__Bounded ? __last1 : __first1 + std::min (__n1, __n_out);
37653762 auto __n1_tmp = __last1_tmp - __first1;
37663763
3767- auto __last2_tmp = !__Bounded ? __last2 : __first2 + std::min (__n2, __n_out > __n1_tmp ? __n_out - __n1_tmp : 0 );
3764+ auto __last2_tmp =
3765+ !__Bounded ? __last2 : __first2 + std::min (__n2, __n_out > __n1_tmp ? __n_out - __n1_tmp : 0 );
37683766 auto __n2_tmp = __last2_tmp - __first2;
37693767
37703768 // {1} < {2}: seq2 is wholly greater than seq1, so, do parallel copying seq1 and seq2
@@ -3774,7 +3772,8 @@ __parallel_set_union_op(__parallel_tag<_IsVector> __tag, _ExecutionPolicy&& __ex
37743772 __internal::__pattern_walk2_brick (__tag, __exec, __first1, __last1_tmp, __result1, __copy_range);
37753773 },
37763774 [=, &__exec] {
3777- __internal::__pattern_walk2_brick (__tag, __exec, __first2, __last2_tmp, __result1 + __n1_tmp, __copy_range);
3775+ __internal::__pattern_walk2_brick (__tag, __exec, __first2, __last2_tmp, __result1 + __n1_tmp,
3776+ __copy_range);
37783777 });
37793778
37803779 return {__last1_tmp, __last2_tmp, __result1 + __n1_tmp + __n2_tmp};
@@ -3790,7 +3789,8 @@ __parallel_set_union_op(__parallel_tag<_IsVector> __tag, _ExecutionPolicy&& __ex
37903789 auto __last2_tmp = !__Bounded ? __last2 : __first2 + std::min (__n2, __n_out);
37913790 auto __n2_tmp = __last2_tmp - __first2;
37923791
3793- auto __last1_tmp = !__Bounded ? __last1 : __first1 + std::min (__n1, __n_out > __n2_tmp ? __n_out - __n2_tmp : 0 );
3792+ auto __last1_tmp =
3793+ !__Bounded ? __last1 : __first1 + std::min (__n1, __n_out > __n2_tmp ? __n_out - __n2_tmp : 0 );
37943794 auto __n1_tmp = __last1_tmp - __first1;
37953795
37963796 // {2} < {1}: seq2 is wholly greater than seq1, so, do parallel copying seq1 and seq2
@@ -3800,7 +3800,8 @@ __parallel_set_union_op(__parallel_tag<_IsVector> __tag, _ExecutionPolicy&& __ex
38003800 __internal::__pattern_walk2_brick (__tag, __exec, __first2, __last2_tmp, __result1, __copy_range);
38013801 },
38023802 [=, &__exec] {
3803- __internal::__pattern_walk2_brick (__tag, __exec, __first1, __last1_tmp, __result1 + __n2_tmp, __copy_range);
3803+ __internal::__pattern_walk2_brick (__tag, __exec, __first1, __last1_tmp, __result1 + __n2_tmp,
3804+ __copy_range);
38043805 });
38053806
38063807 return {__last1_tmp, __last2_tmp, __result1 + __n1_tmp + __n2_tmp};
@@ -3811,18 +3812,18 @@ __parallel_set_union_op(__parallel_tag<_IsVector> __tag, _ExecutionPolicy&& __ex
38113812 const auto __m1 = __left_bound_seq_1 - __first1;
38123813 if (oneapi::dpl::__internal::__is_great_that_set_algo_cut_off (__m1))
38133814 {
3814- oneapi::dpl::__utils::__set_operations_result<_RandomAccessIterator1, _RandomAccessIterator2, _OutputIterator> __finish;
3815+ oneapi::dpl::__utils::__set_operations_result<_RandomAccessIterator1, _RandomAccessIterator2, _OutputIterator>
3816+ __finish;
38153817
38163818 auto __res_or = __result1;
3817- __result1 += __m1; // we know proper offset due to [first1; left_bound_seq_1) < [first2; last2)
3819+ __result1 += __m1; // we know proper offset due to [first1; left_bound_seq_1) < [first2; last2)
38183820 __par_backend::__parallel_invoke (
38193821 __backend_tag{}, __exec,
38203822 // do parallel copying of [first1; left_bound_seq_1)
38213823 [=, &__exec] {
38223824 __internal::__pattern_walk2_brick (__tag, __exec, __first1, __left_bound_seq_1, __res_or, __copy_range);
38233825 },
3824- [=, &__exec, &__finish]
3825- {
3826+ [=, &__exec, &__finish] {
38263827 __finish = __internal::__parallel_set_op<__Bounded>(
38273828 __tag, __exec, __left_bound_seq_1, __last1, __first2, __last2, __result1, __result2, __size_fnc,
38283829 __set_union_op, __reached_positions_evaluator, __comp, __proj1, __proj2);
@@ -3834,18 +3835,18 @@ __parallel_set_union_op(__parallel_tag<_IsVector> __tag, _ExecutionPolicy&& __ex
38343835 assert (__m1 == 0 || __m2 == 0 );
38353836 if (oneapi::dpl::__internal::__is_great_that_set_algo_cut_off (__m2))
38363837 {
3837- oneapi::dpl::__utils::__set_operations_result<_RandomAccessIterator1, _RandomAccessIterator2, _OutputIterator> __finish;
3838+ oneapi::dpl::__utils::__set_operations_result<_RandomAccessIterator1, _RandomAccessIterator2, _OutputIterator>
3839+ __finish;
38383840
38393841 auto __res_or = __result1;
3840- __result1 += __m2; // we know proper offset due to [first2; left_bound_seq_2) < [first1; last1)
3842+ __result1 += __m2; // we know proper offset due to [first2; left_bound_seq_2) < [first1; last1)
38413843 __par_backend::__parallel_invoke (
38423844 __backend_tag{}, __exec,
38433845 // do parallel copying of [first2; left_bound_seq_2)
38443846 [=, &__exec] {
38453847 __internal::__pattern_walk2_brick (__tag, __exec, __first2, __left_bound_seq_2, __res_or, __copy_range);
38463848 },
3847- [=, &__exec, &__finish]
3848- {
3849+ [=, &__exec, &__finish] {
38493850 __finish = __internal::__parallel_set_op<__Bounded>(
38503851 __tag, __exec, __first1, __last1, __left_bound_seq_2, __last2, __result1, __result2, __size_fnc,
38513852 __set_union_op, __reached_positions_evaluator, __comp, __proj1, __proj2);
@@ -3920,7 +3921,7 @@ __pattern_set_union(__parallel_tag<_IsVector> __tag, _ExecutionPolicy&& __exec,
39203921 return std::set_union (__first1, __last1, __first2, __last2, __result, __comp);
39213922
39223923 using _Tp = typename std::iterator_traits<_OutputIterator>::value_type;
3923- return __parallel_set_union_op</* __Bounded*/ false >(
3924+ return __parallel_set_union_op</* __Bounded*/ false >(
39243925 __tag, std::forward<_ExecutionPolicy>(__exec), __first1, __last1, __first2, __last2, __result,
39253926 __result + __n1 + __n2,
39263927 [](_RandomAccessIterator1 __first1, _RandomAccessIterator1 __last1, _RandomAccessIterator2 __first2,
0 commit comments