@@ -676,42 +676,42 @@ macro_rules! atomic_int {
676
676
impl AtomicOperations for $int_type {
677
677
#[ inline( always) ]
678
678
unsafe fn atomic_store( dst: * mut Self , val: Self ) {
679
- :: core:: intrinsics:: atomic_store ( dst, val) ;
679
+ :: core:: intrinsics:: atomic_store_seqcst ( dst, val) ;
680
680
}
681
681
682
682
#[ inline( always) ]
683
683
unsafe fn atomic_load( dst: * const Self ) -> Self {
684
- :: core:: intrinsics:: atomic_load ( dst)
684
+ :: core:: intrinsics:: atomic_load_seqcst ( dst)
685
685
}
686
686
687
687
#[ inline( always) ]
688
688
unsafe fn atomic_add( dst: * mut Self , val: Self ) {
689
- :: core:: intrinsics:: atomic_xadd ( dst, val) ;
689
+ :: core:: intrinsics:: atomic_xadd_seqcst ( dst, val) ;
690
690
}
691
691
692
692
#[ inline( always) ]
693
693
unsafe fn atomic_sub( dst: * mut Self , val: Self ) {
694
- :: core:: intrinsics:: atomic_xsub ( dst, val) ;
694
+ :: core:: intrinsics:: atomic_xsub_seqcst ( dst, val) ;
695
695
}
696
696
697
697
#[ inline( always) ]
698
698
unsafe fn atomic_and( dst: * mut Self , val: Self ) {
699
- :: core:: intrinsics:: atomic_and ( dst, val) ;
699
+ :: core:: intrinsics:: atomic_and_seqcst ( dst, val) ;
700
700
}
701
701
702
702
#[ inline( always) ]
703
703
unsafe fn atomic_clear( dst: * mut Self , val: Self ) {
704
- :: core:: intrinsics:: atomic_and ( dst, !val) ;
704
+ :: core:: intrinsics:: atomic_and_seqcst ( dst, !val) ;
705
705
}
706
706
707
707
#[ inline( always) ]
708
708
unsafe fn atomic_or( dst: * mut Self , val: Self ) {
709
- :: core:: intrinsics:: atomic_or ( dst, val) ;
709
+ :: core:: intrinsics:: atomic_or_seqcst ( dst, val) ;
710
710
}
711
711
712
712
#[ inline( always) ]
713
713
unsafe fn atomic_xor( dst: * mut Self , val: Self ) {
714
- :: core:: intrinsics:: atomic_xor ( dst, val) ;
714
+ :: core:: intrinsics:: atomic_xor_seqcst ( dst, val) ;
715
715
}
716
716
}
717
717
}
0 commit comments