Lines Matching refs:__result
406 __bit_iterator<_Cp, false> __result)
423 *__result.__seg_ &= ~__m;
424 *__result.__seg_ |= __b;
425 __result.__seg_ += (__dn + __result.__ctz_) / __bits_per_word;
426 __result.__ctz_ = static_cast<unsigned>((__dn + __result.__ctz_) % __bits_per_word);
433 _VSTD::memmove(_VSTD::__to_raw_pointer(__result.__seg_),
437 __result.__seg_ += __nw;
444 *__result.__seg_ &= ~__m;
445 *__result.__seg_ |= __b;
446 __result.__ctz_ = static_cast<unsigned>(__n);
449 return __result;
455 __bit_iterator<_Cp, false> __result)
472 unsigned __clz_r = __bits_per_word - __result.__ctz_;
474 … __m = (~__storage_type(0) << __result.__ctz_) & (~__storage_type(0) >> (__clz_r - __ddn));
475 *__result.__seg_ &= ~__m;
476 if (__result.__ctz_ > __first.__ctz_)
477 *__result.__seg_ |= __b << (__result.__ctz_ - __first.__ctz_);
479 *__result.__seg_ |= __b >> (__first.__ctz_ - __result.__ctz_);
480 __result.__seg_ += (__ddn + __result.__ctz_) / __bits_per_word;
481 __result.__ctz_ = static_cast<unsigned>((__ddn + __result.__ctz_) % __bits_per_word);
486 *__result.__seg_ &= ~__m;
487 *__result.__seg_ |= __b >> (__first.__ctz_ + __ddn);
488 __result.__ctz_ = static_cast<unsigned>(__dn);
495 unsigned __clz_r = __bits_per_word - __result.__ctz_;
496 __storage_type __m = ~__storage_type(0) << __result.__ctz_;
500 *__result.__seg_ &= ~__m;
501 *__result.__seg_ |= __b << __result.__ctz_;
502 ++__result.__seg_;
503 *__result.__seg_ &= __m;
504 *__result.__seg_ |= __b >> __clz_r;
512 … __m = (~__storage_type(0) << __result.__ctz_) & (~__storage_type(0) >> (__clz_r - __dn));
513 *__result.__seg_ &= ~__m;
514 *__result.__seg_ |= __b << __result.__ctz_;
515 __result.__seg_ += (__dn + __result.__ctz_) / __bits_per_word;
516 __result.__ctz_ = static_cast<unsigned>((__dn + __result.__ctz_) % __bits_per_word);
521 *__result.__seg_ &= ~__m;
522 *__result.__seg_ |= __b >> __dn;
523 __result.__ctz_ = static_cast<unsigned>(__n);
527 return __result;
533 …<_Cp, _IsConst> __first, __bit_iterator<_Cp, _IsConst> __last, __bit_iterator<_Cp, false> __result)
535 if (__first.__ctz_ == __result.__ctz_)
536 return __copy_aligned(__first, __last, __result);
537 return __copy_unaligned(__first, __last, __result);
545 __bit_iterator<_Cp, false> __result)
562 *__result.__seg_ &= ~__m;
563 *__result.__seg_ |= __b;
564 __result.__ctz_ = static_cast<unsigned>(((-__dn & (__bits_per_word - 1)) +
565 __result.__ctz_) % __bits_per_word);
569 // __result.__ctz_ == 0 || __n == 0
572 __result.__seg_ -= __nw;
574 _VSTD::memmove(_VSTD::__to_raw_pointer(__result.__seg_),
583 *--__result.__seg_ &= ~__m;
584 *__result.__seg_ |= __b;
585 __result.__ctz_ = static_cast<unsigned>(-__n & (__bits_per_word - 1));
588 return __result;
594 __bit_iterator<_Cp, false> __result)
611 unsigned __clz_r = __bits_per_word - __result.__ctz_;
612 __storage_type __ddn = _VSTD::min(__dn, static_cast<difference_type>(__result.__ctz_));
615 … __m = (~__storage_type(0) << (__result.__ctz_ - __ddn)) & (~__storage_type(0) >> __clz_r);
616 *__result.__seg_ &= ~__m;
617 if (__result.__ctz_ > __last.__ctz_)
618 *__result.__seg_ |= __b << (__result.__ctz_ - __last.__ctz_);
620 *__result.__seg_ |= __b >> (__last.__ctz_ - __result.__ctz_);
621 __result.__ctz_ = static_cast<unsigned>(((-__ddn & (__bits_per_word - 1)) +
622 __result.__ctz_) % __bits_per_word);
627 // __result.__ctz_ == 0
628 --__result.__seg_;
629 __result.__ctz_ = static_cast<unsigned>(-__dn & (__bits_per_word - 1));
630 __m = ~__storage_type(0) << __result.__ctz_;
631 *__result.__seg_ &= ~__m;
633 *__result.__seg_ |= __b << (__result.__ctz_ - __last.__ctz_);
638 // __result.__ctz_ != 0 || __n == 0
640 unsigned __clz_r = __bits_per_word - __result.__ctz_;
645 *__result.__seg_ &= ~__m;
646 *__result.__seg_ |= __b >> __clz_r;
647 *--__result.__seg_ &= __m;
648 *__result.__seg_ |= __b << __result.__ctz_;
655 __clz_r = __bits_per_word - __result.__ctz_;
656 __storage_type __dn = _VSTD::min(__n, static_cast<difference_type>(__result.__ctz_));
657 … __m = (~__storage_type(0) << (__result.__ctz_ - __dn)) & (~__storage_type(0) >> __clz_r);
658 *__result.__seg_ &= ~__m;
659 *__result.__seg_ |= __b >> (__bits_per_word - __result.__ctz_);
660 __result.__ctz_ = static_cast<unsigned>(((-__dn & (__bits_per_word - 1)) +
661 __result.__ctz_) % __bits_per_word);
665 // __result.__ctz_ == 0
666 --__result.__seg_;
667 __result.__ctz_ = static_cast<unsigned>(-__n & (__bits_per_word - 1));
668 __m = ~__storage_type(0) << __result.__ctz_;
669 *__result.__seg_ &= ~__m;
670 *__result.__seg_ |= __b << (__result.__ctz_ - (__bits_per_word - __n - __dn));
674 return __result;
680 …<_Cp, _IsConst> __first, __bit_iterator<_Cp, _IsConst> __last, __bit_iterator<_Cp, false> __result)
682 if (__last.__ctz_ == __result.__ctz_)
683 return __copy_backward_aligned(__first, __last, __result);
684 return __copy_backward_unaligned(__first, __last, __result);
692 …<_Cp, _IsConst> __first, __bit_iterator<_Cp, _IsConst> __last, __bit_iterator<_Cp, false> __result)
694 return _VSTD::copy(__first, __last, __result);
702 …<_Cp, _IsConst> __first, __bit_iterator<_Cp, _IsConst> __last, __bit_iterator<_Cp, false> __result)
704 return _VSTD::copy_backward(__first, __last, __result);
712 __bit_iterator<__C2, false> __result)
730 __storage_type __b2 = *__result.__seg_ & __m;
731 *__result.__seg_ &= ~__m;
732 *__result.__seg_ |= __b1;
734 __result.__seg_ += (__dn + __result.__ctz_) / __bits_per_word;
735 __result.__ctz_ = static_cast<unsigned>((__dn + __result.__ctz_) % __bits_per_word);
741 for (; __n >= __bits_per_word; __n -= __bits_per_word, ++__first.__seg_, ++__result.__seg_)
742 swap(*__first.__seg_, *__result.__seg_);
749 __storage_type __b2 = *__result.__seg_ & __m;
750 *__result.__seg_ &= ~__m;
751 *__result.__seg_ |= __b1;
753 __result.__ctz_ = static_cast<unsigned>(__n);
756 return __result;
762 __bit_iterator<__C2, false> __result)
780 unsigned __clz_r = __bits_per_word - __result.__ctz_;
782 … __m = (~__storage_type(0) << __result.__ctz_) & (~__storage_type(0) >> (__clz_r - __ddn));
783 __storage_type __b2 = *__result.__seg_ & __m;
784 *__result.__seg_ &= ~__m;
785 if (__result.__ctz_ > __first.__ctz_)
787 unsigned __s = __result.__ctz_ - __first.__ctz_;
788 *__result.__seg_ |= __b1 << __s;
793 unsigned __s = __first.__ctz_ - __result.__ctz_;
794 *__result.__seg_ |= __b1 >> __s;
797 __result.__seg_ += (__ddn + __result.__ctz_) / __bits_per_word;
798 __result.__ctz_ = static_cast<unsigned>((__ddn + __result.__ctz_) % __bits_per_word);
803 __b2 = *__result.__seg_ & __m;
804 *__result.__seg_ &= ~__m;
806 *__result.__seg_ |= __b1 >> __s;
808 __result.__ctz_ = static_cast<unsigned>(__dn);
815 __storage_type __m = ~__storage_type(0) << __result.__ctz_;
816 unsigned __clz_r = __bits_per_word - __result.__ctz_;
820 __storage_type __b2 = *__result.__seg_ & __m;
821 *__result.__seg_ &= ~__m;
822 *__result.__seg_ |= __b1 << __result.__ctz_;
823 *__first.__seg_ = __b2 >> __result.__ctz_;
824 ++__result.__seg_;
825 __b2 = *__result.__seg_ & ~__m;
826 *__result.__seg_ &= __m;
827 *__result.__seg_ |= __b1 >> __clz_r;
837 … __m = (~__storage_type(0) << __result.__ctz_) & (~__storage_type(0) >> (__clz_r - __dn));
838 __storage_type __b2 = *__result.__seg_ & __m;
839 *__result.__seg_ &= ~__m;
840 *__result.__seg_ |= __b1 << __result.__ctz_;
841 *__first.__seg_ |= __b2 >> __result.__ctz_;
842 __result.__seg_ += (__dn + __result.__ctz_) / __bits_per_word;
843 __result.__ctz_ = static_cast<unsigned>((__dn + __result.__ctz_) % __bits_per_word);
848 __b2 = *__result.__seg_ & __m;
849 *__result.__seg_ &= ~__m;
850 *__result.__seg_ |= __b1 >> __dn;
852 __result.__ctz_ = static_cast<unsigned>(__n);
856 return __result;
1230 … __bit_iterator<_Dp, false> __result);
1233 … __bit_iterator<_Dp, false> __result);
1236 … __bit_iterator<_Dp, false> __result);
1239 … __bit_iterator<_Dp, false> __result);
1242 … __bit_iterator<_Dp, false> __result);
1245 … __bit_iterator<_Dp, false> __result);