Struct scale_info::prelude::sync::atomic::AtomicUsize
1.0.0 · source · #[repr(C, align(8))]pub struct AtomicUsize { /* private fields */ }
Expand description
An integer type which can be safely shared between threads.
This type has the same
size and bit validity
as the underlying integer type, usize
.
However, the alignment of this type is always equal to its size, even on targets where usize
has a lesser alignment.
For more about the differences between atomic types and non-atomic types as well as information about the portability of this type, please see the module-level documentation.
Note: This type is only available on platforms that support
atomic loads and stores of usize
.
Implementations§
source§impl AtomicUsize
impl AtomicUsize
1.0.0 (const: 1.24.0) · sourcepub const fn new(v: usize) -> AtomicUsize
pub const fn new(v: usize) -> AtomicUsize
Creates a new atomic integer.
§Examples
use std::sync::atomic::AtomicUsize;
let atomic_forty_two = AtomicUsize::new(42);
1.75.0 (const: unstable) · sourcepub unsafe fn from_ptr<'a>(ptr: *mut usize) -> &'a AtomicUsize
pub unsafe fn from_ptr<'a>(ptr: *mut usize) -> &'a AtomicUsize
Creates a new reference to an atomic integer from a pointer.
§Examples
use std::sync::atomic::{self, AtomicUsize};
// Get a pointer to an allocated value
let ptr: *mut usize = Box::into_raw(Box::new(0));
assert!(ptr.cast::<AtomicUsize>().is_aligned());
{
// Create an atomic view of the allocated value
let atomic = unsafe {AtomicUsize::from_ptr(ptr) };
// Use `atomic` for atomic operations, possibly share it with other threads
atomic.store(1, atomic::Ordering::Relaxed);
}
// It's ok to non-atomically access the value behind `ptr`,
// since the reference to the atomic ended its lifetime in the block above
assert_eq!(unsafe { *ptr }, 1);
// Deallocate the value
unsafe { drop(Box::from_raw(ptr)) }
§Safety
ptr
must be aligned toalign_of::<AtomicUsize>()
(note that on some platforms this can be bigger thanalign_of::<usize>()
).ptr
must be valid for both reads and writes for the whole lifetime'a
.- You must adhere to the Memory model for atomic accesses. In particular, it is not allowed to mix atomic and non-atomic accesses, or atomic accesses of different sizes, without synchronization.
1.15.0 · sourcepub fn get_mut(&mut self) -> &mut usize
pub fn get_mut(&mut self) -> &mut usize
Returns a mutable reference to the underlying integer.
This is safe because the mutable reference guarantees that no other threads are concurrently accessing the atomic data.
§Examples
use std::sync::atomic::{AtomicUsize, Ordering};
let mut some_var = AtomicUsize::new(10);
assert_eq!(*some_var.get_mut(), 10);
*some_var.get_mut() = 5;
assert_eq!(some_var.load(Ordering::SeqCst), 5);
sourcepub fn from_mut(v: &mut usize) -> &mut AtomicUsize
🔬This is a nightly-only experimental API. (atomic_from_mut
)
pub fn from_mut(v: &mut usize) -> &mut AtomicUsize
atomic_from_mut
)Get atomic access to a &mut usize
.
Note: This function is only available on targets where usize
has an alignment of 8 bytes.
§Examples
#![feature(atomic_from_mut)]
use std::sync::atomic::{AtomicUsize, Ordering};
let mut some_int = 123;
let a = AtomicUsize::from_mut(&mut some_int);
a.store(100, Ordering::Relaxed);
assert_eq!(some_int, 100);
sourcepub fn get_mut_slice(this: &mut [AtomicUsize]) -> &mut [usize]
🔬This is a nightly-only experimental API. (atomic_from_mut
)
pub fn get_mut_slice(this: &mut [AtomicUsize]) -> &mut [usize]
atomic_from_mut
)Get non-atomic access to a &mut [AtomicUsize]
slice
This is safe because the mutable reference guarantees that no other threads are concurrently accessing the atomic data.
§Examples
#![feature(atomic_from_mut)]
use std::sync::atomic::{AtomicUsize, Ordering};
let mut some_ints = [const { AtomicUsize::new(0) }; 10];
let view: &mut [usize] = AtomicUsize::get_mut_slice(&mut some_ints);
assert_eq!(view, [0; 10]);
view
.iter_mut()
.enumerate()
.for_each(|(idx, int)| *int = idx as _);
std::thread::scope(|s| {
some_ints
.iter()
.enumerate()
.for_each(|(idx, int)| {
s.spawn(move || assert_eq!(int.load(Ordering::Relaxed), idx as _));
})
});
sourcepub fn from_mut_slice(v: &mut [usize]) -> &mut [AtomicUsize]
🔬This is a nightly-only experimental API. (atomic_from_mut
)
pub fn from_mut_slice(v: &mut [usize]) -> &mut [AtomicUsize]
atomic_from_mut
)Get atomic access to a &mut [usize]
slice.
§Examples
#![feature(atomic_from_mut)]
use std::sync::atomic::{AtomicUsize, Ordering};
let mut some_ints = [0; 10];
let a = &*AtomicUsize::from_mut_slice(&mut some_ints);
std::thread::scope(|s| {
for i in 0..a.len() {
s.spawn(move || a[i].store(i as _, Ordering::Relaxed));
}
});
for (i, n) in some_ints.into_iter().enumerate() {
assert_eq!(i, n as usize);
}
1.15.0 (const: 1.79.0) · sourcepub const fn into_inner(self) -> usize
pub const fn into_inner(self) -> usize
Consumes the atomic and returns the contained value.
This is safe because passing self
by value guarantees that no other threads are
concurrently accessing the atomic data.
§Examples
use std::sync::atomic::AtomicUsize;
let some_var = AtomicUsize::new(5);
assert_eq!(some_var.into_inner(), 5);
1.0.0 · sourcepub fn load(&self, order: Ordering) -> usize
pub fn load(&self, order: Ordering) -> usize
Loads a value from the atomic integer.
load
takes an Ordering
argument which describes the memory ordering of this operation.
Possible values are SeqCst
, Acquire
and Relaxed
.
§Panics
Panics if order
is Release
or AcqRel
.
§Examples
use std::sync::atomic::{AtomicUsize, Ordering};
let some_var = AtomicUsize::new(5);
assert_eq!(some_var.load(Ordering::Relaxed), 5);
1.0.0 · sourcepub fn store(&self, val: usize, order: Ordering)
pub fn store(&self, val: usize, order: Ordering)
Stores a value into the atomic integer.
store
takes an Ordering
argument which describes the memory ordering of this operation.
Possible values are SeqCst
, Release
and Relaxed
.
§Panics
Panics if order
is Acquire
or AcqRel
.
§Examples
use std::sync::atomic::{AtomicUsize, Ordering};
let some_var = AtomicUsize::new(5);
some_var.store(10, Ordering::Relaxed);
assert_eq!(some_var.load(Ordering::Relaxed), 10);
1.0.0 · sourcepub fn swap(&self, val: usize, order: Ordering) -> usize
pub fn swap(&self, val: usize, order: Ordering) -> usize
Stores a value into the atomic integer, returning the previous value.
swap
takes an Ordering
argument which describes the memory ordering
of this operation. All ordering modes are possible. Note that using
Acquire
makes the store part of this operation Relaxed
, and
using Release
makes the load part Relaxed
.
Note: This method is only available on platforms that support atomic operations on
usize
.
§Examples
use std::sync::atomic::{AtomicUsize, Ordering};
let some_var = AtomicUsize::new(5);
assert_eq!(some_var.swap(10, Ordering::Relaxed), 5);
1.0.0 · sourcepub fn compare_and_swap(
&self,
current: usize,
new: usize,
order: Ordering,
) -> usize
👎Deprecated since 1.50.0: Use compare_exchange
or compare_exchange_weak
instead
pub fn compare_and_swap( &self, current: usize, new: usize, order: Ordering, ) -> usize
compare_exchange
or compare_exchange_weak
insteadStores a value into the atomic integer if the current value is the same as
the current
value.
The return value is always the previous value. If it is equal to current
, then the
value was updated.
compare_and_swap
also takes an Ordering
argument which describes the memory
ordering of this operation. Notice that even when using AcqRel
, the operation
might fail and hence just perform an Acquire
load, but not have Release
semantics.
Using Acquire
makes the store part of this operation Relaxed
if it
happens, and using Release
makes the load part Relaxed
.
Note: This method is only available on platforms that support atomic operations on
usize
.
§Migrating to compare_exchange
and compare_exchange_weak
compare_and_swap
is equivalent to compare_exchange
with the following mapping for
memory orderings:
Original | Success | Failure |
---|---|---|
Relaxed | Relaxed | Relaxed |
Acquire | Acquire | Acquire |
Release | Release | Relaxed |
AcqRel | AcqRel | Acquire |
SeqCst | SeqCst | SeqCst |
compare_exchange_weak
is allowed to fail spuriously even when the comparison succeeds,
which allows the compiler to generate better assembly code when the compare and swap
is used in a loop.
§Examples
use std::sync::atomic::{AtomicUsize, Ordering};
let some_var = AtomicUsize::new(5);
assert_eq!(some_var.compare_and_swap(5, 10, Ordering::Relaxed), 5);
assert_eq!(some_var.load(Ordering::Relaxed), 10);
assert_eq!(some_var.compare_and_swap(6, 12, Ordering::Relaxed), 10);
assert_eq!(some_var.load(Ordering::Relaxed), 10);
1.10.0 · sourcepub fn compare_exchange(
&self,
current: usize,
new: usize,
success: Ordering,
failure: Ordering,
) -> Result<usize, usize>
pub fn compare_exchange( &self, current: usize, new: usize, success: Ordering, failure: Ordering, ) -> Result<usize, usize>
Stores a value into the atomic integer if the current value is the same as
the current
value.
The return value is a result indicating whether the new value was written and
containing the previous value. On success this value is guaranteed to be equal to
current
.
compare_exchange
takes two Ordering
arguments to describe the memory
ordering of this operation. success
describes the required ordering for the
read-modify-write operation that takes place if the comparison with current
succeeds.
failure
describes the required ordering for the load operation that takes place when
the comparison fails. Using Acquire
as success ordering makes the store part
of this operation Relaxed
, and using Release
makes the successful load
Relaxed
. The failure ordering can only be SeqCst
, Acquire
or Relaxed
.
Note: This method is only available on platforms that support atomic operations on
usize
.
§Examples
use std::sync::atomic::{AtomicUsize, Ordering};
let some_var = AtomicUsize::new(5);
assert_eq!(some_var.compare_exchange(5, 10,
Ordering::Acquire,
Ordering::Relaxed),
Ok(5));
assert_eq!(some_var.load(Ordering::Relaxed), 10);
assert_eq!(some_var.compare_exchange(6, 12,
Ordering::SeqCst,
Ordering::Acquire),
Err(10));
assert_eq!(some_var.load(Ordering::Relaxed), 10);
1.10.0 · sourcepub fn compare_exchange_weak(
&self,
current: usize,
new: usize,
success: Ordering,
failure: Ordering,
) -> Result<usize, usize>
pub fn compare_exchange_weak( &self, current: usize, new: usize, success: Ordering, failure: Ordering, ) -> Result<usize, usize>
Stores a value into the atomic integer if the current value is the same as
the current
value.
Unlike AtomicUsize::compare_exchange
,
this function is allowed to spuriously fail even
when the comparison succeeds, which can result in more efficient code on some
platforms. The return value is a result indicating whether the new value was
written and containing the previous value.
compare_exchange_weak
takes two Ordering
arguments to describe the memory
ordering of this operation. success
describes the required ordering for the
read-modify-write operation that takes place if the comparison with current
succeeds.
failure
describes the required ordering for the load operation that takes place when
the comparison fails. Using Acquire
as success ordering makes the store part
of this operation Relaxed
, and using Release
makes the successful load
Relaxed
. The failure ordering can only be SeqCst
, Acquire
or Relaxed
.
Note: This method is only available on platforms that support atomic operations on
usize
.
§Examples
use std::sync::atomic::{AtomicUsize, Ordering};
let val = AtomicUsize::new(4);
let mut old = val.load(Ordering::Relaxed);
loop {
let new = old * 2;
match val.compare_exchange_weak(old, new, Ordering::SeqCst, Ordering::Relaxed) {
Ok(_) => break,
Err(x) => old = x,
}
}
1.0.0 · sourcepub fn fetch_add(&self, val: usize, order: Ordering) -> usize
pub fn fetch_add(&self, val: usize, order: Ordering) -> usize
Adds to the current value, returning the previous value.
This operation wraps around on overflow.
fetch_add
takes an Ordering
argument which describes the memory ordering
of this operation. All ordering modes are possible. Note that using
Acquire
makes the store part of this operation Relaxed
, and
using Release
makes the load part Relaxed
.
Note: This method is only available on platforms that support atomic operations on
usize
.
§Examples
use std::sync::atomic::{AtomicUsize, Ordering};
let foo = AtomicUsize::new(0);
assert_eq!(foo.fetch_add(10, Ordering::SeqCst), 0);
assert_eq!(foo.load(Ordering::SeqCst), 10);
1.0.0 · sourcepub fn fetch_sub(&self, val: usize, order: Ordering) -> usize
pub fn fetch_sub(&self, val: usize, order: Ordering) -> usize
Subtracts from the current value, returning the previous value.
This operation wraps around on overflow.
fetch_sub
takes an Ordering
argument which describes the memory ordering
of this operation. All ordering modes are possible. Note that using
Acquire
makes the store part of this operation Relaxed
, and
using Release
makes the load part Relaxed
.
Note: This method is only available on platforms that support atomic operations on
usize
.
§Examples
use std::sync::atomic::{AtomicUsize, Ordering};
let foo = AtomicUsize::new(20);
assert_eq!(foo.fetch_sub(10, Ordering::SeqCst), 20);
assert_eq!(foo.load(Ordering::SeqCst), 10);
1.0.0 · sourcepub fn fetch_and(&self, val: usize, order: Ordering) -> usize
pub fn fetch_and(&self, val: usize, order: Ordering) -> usize
Bitwise “and” with the current value.
Performs a bitwise “and” operation on the current value and the argument val
, and
sets the new value to the result.
Returns the previous value.
fetch_and
takes an Ordering
argument which describes the memory ordering
of this operation. All ordering modes are possible. Note that using
Acquire
makes the store part of this operation Relaxed
, and
using Release
makes the load part Relaxed
.
Note: This method is only available on platforms that support atomic operations on
usize
.
§Examples
use std::sync::atomic::{AtomicUsize, Ordering};
let foo = AtomicUsize::new(0b101101);
assert_eq!(foo.fetch_and(0b110011, Ordering::SeqCst), 0b101101);
assert_eq!(foo.load(Ordering::SeqCst), 0b100001);
1.27.0 · sourcepub fn fetch_nand(&self, val: usize, order: Ordering) -> usize
pub fn fetch_nand(&self, val: usize, order: Ordering) -> usize
Bitwise “nand” with the current value.
Performs a bitwise “nand” operation on the current value and the argument val
, and
sets the new value to the result.
Returns the previous value.
fetch_nand
takes an Ordering
argument which describes the memory ordering
of this operation. All ordering modes are possible. Note that using
Acquire
makes the store part of this operation Relaxed
, and
using Release
makes the load part Relaxed
.
Note: This method is only available on platforms that support atomic operations on
usize
.
§Examples
use std::sync::atomic::{AtomicUsize, Ordering};
let foo = AtomicUsize::new(0x13);
assert_eq!(foo.fetch_nand(0x31, Ordering::SeqCst), 0x13);
assert_eq!(foo.load(Ordering::SeqCst), !(0x13 & 0x31));
1.0.0 · sourcepub fn fetch_or(&self, val: usize, order: Ordering) -> usize
pub fn fetch_or(&self, val: usize, order: Ordering) -> usize
Bitwise “or” with the current value.
Performs a bitwise “or” operation on the current value and the argument val
, and
sets the new value to the result.
Returns the previous value.
fetch_or
takes an Ordering
argument which describes the memory ordering
of this operation. All ordering modes are possible. Note that using
Acquire
makes the store part of this operation Relaxed
, and
using Release
makes the load part Relaxed
.
Note: This method is only available on platforms that support atomic operations on
usize
.
§Examples
use std::sync::atomic::{AtomicUsize, Ordering};
let foo = AtomicUsize::new(0b101101);
assert_eq!(foo.fetch_or(0b110011, Ordering::SeqCst), 0b101101);
assert_eq!(foo.load(Ordering::SeqCst), 0b111111);
1.0.0 · sourcepub fn fetch_xor(&self, val: usize, order: Ordering) -> usize
pub fn fetch_xor(&self, val: usize, order: Ordering) -> usize
Bitwise “xor” with the current value.
Performs a bitwise “xor” operation on the current value and the argument val
, and
sets the new value to the result.
Returns the previous value.
fetch_xor
takes an Ordering
argument which describes the memory ordering
of this operation. All ordering modes are possible. Note that using
Acquire
makes the store part of this operation Relaxed
, and
using Release
makes the load part Relaxed
.
Note: This method is only available on platforms that support atomic operations on
usize
.
§Examples
use std::sync::atomic::{AtomicUsize, Ordering};
let foo = AtomicUsize::new(0b101101);
assert_eq!(foo.fetch_xor(0b110011, Ordering::SeqCst), 0b101101);
assert_eq!(foo.load(Ordering::SeqCst), 0b011110);
1.45.0 · sourcepub fn fetch_update<F>(
&self,
set_order: Ordering,
fetch_order: Ordering,
f: F,
) -> Result<usize, usize>
pub fn fetch_update<F>( &self, set_order: Ordering, fetch_order: Ordering, f: F, ) -> Result<usize, usize>
Fetches the value, and applies a function to it that returns an optional
new value. Returns a Result
of Ok(previous_value)
if the function returned Some(_)
, else
Err(previous_value)
.
Note: This may call the function multiple times if the value has been changed from other threads in
the meantime, as long as the function returns Some(_)
, but the function will have been applied
only once to the stored value.
fetch_update
takes two Ordering
arguments to describe the memory ordering of this operation.
The first describes the required ordering for when the operation finally succeeds while the second
describes the required ordering for loads. These correspond to the success and failure orderings of
AtomicUsize::compare_exchange
respectively.
Using Acquire
as success ordering makes the store part
of this operation Relaxed
, and using Release
makes the final successful load
Relaxed
. The (failed) load ordering can only be SeqCst
, Acquire
or Relaxed
.
Note: This method is only available on platforms that support atomic operations on
usize
.
§Considerations
This method is not magic; it is not provided by the hardware.
It is implemented in terms of
AtomicUsize::compare_exchange_weak
,
and suffers from the same drawbacks.
In particular, this method will not circumvent the ABA Problem.
§Examples
use std::sync::atomic::{AtomicUsize, Ordering};
let x = AtomicUsize::new(7);
assert_eq!(x.fetch_update(Ordering::SeqCst, Ordering::SeqCst, |_| None), Err(7));
assert_eq!(x.fetch_update(Ordering::SeqCst, Ordering::SeqCst, |x| Some(x + 1)), Ok(7));
assert_eq!(x.fetch_update(Ordering::SeqCst, Ordering::SeqCst, |x| Some(x + 1)), Ok(8));
assert_eq!(x.load(Ordering::SeqCst), 9);
1.45.0 · sourcepub fn fetch_max(&self, val: usize, order: Ordering) -> usize
pub fn fetch_max(&self, val: usize, order: Ordering) -> usize
Maximum with the current value.
Finds the maximum of the current value and the argument val
, and
sets the new value to the result.
Returns the previous value.
fetch_max
takes an Ordering
argument which describes the memory ordering
of this operation. All ordering modes are possible. Note that using
Acquire
makes the store part of this operation Relaxed
, and
using Release
makes the load part Relaxed
.
Note: This method is only available on platforms that support atomic operations on
usize
.
§Examples
use std::sync::atomic::{AtomicUsize, Ordering};
let foo = AtomicUsize::new(23);
assert_eq!(foo.fetch_max(42, Ordering::SeqCst), 23);
assert_eq!(foo.load(Ordering::SeqCst), 42);
If you want to obtain the maximum value in one step, you can use the following:
use std::sync::atomic::{AtomicUsize, Ordering};
let foo = AtomicUsize::new(23);
let bar = 42;
let max_foo = foo.fetch_max(bar, Ordering::SeqCst).max(bar);
assert!(max_foo == 42);
1.45.0 · sourcepub fn fetch_min(&self, val: usize, order: Ordering) -> usize
pub fn fetch_min(&self, val: usize, order: Ordering) -> usize
Minimum with the current value.
Finds the minimum of the current value and the argument val
, and
sets the new value to the result.
Returns the previous value.
fetch_min
takes an Ordering
argument which describes the memory ordering
of this operation. All ordering modes are possible. Note that using
Acquire
makes the store part of this operation Relaxed
, and
using Release
makes the load part Relaxed
.
Note: This method is only available on platforms that support atomic operations on
usize
.
§Examples
use std::sync::atomic::{AtomicUsize, Ordering};
let foo = AtomicUsize::new(23);
assert_eq!(foo.fetch_min(42, Ordering::Relaxed), 23);
assert_eq!(foo.load(Ordering::Relaxed), 23);
assert_eq!(foo.fetch_min(22, Ordering::Relaxed), 23);
assert_eq!(foo.load(Ordering::Relaxed), 22);
If you want to obtain the minimum value in one step, you can use the following:
use std::sync::atomic::{AtomicUsize, Ordering};
let foo = AtomicUsize::new(23);
let bar = 12;
let min_foo = foo.fetch_min(bar, Ordering::SeqCst).min(bar);
assert_eq!(min_foo, 12);
1.70.0 (const: 1.70.0) · sourcepub const fn as_ptr(&self) -> *mut usize
pub const fn as_ptr(&self) -> *mut usize
Returns a mutable pointer to the underlying integer.
Doing non-atomic reads and writes on the resulting integer can be a data race.
This method is mostly useful for FFI, where the function signature may use
*mut usize
instead of &AtomicUsize
.
Returning an *mut
pointer from a shared reference to this atomic is safe because the
atomic types work with interior mutability. All modifications of an atomic change the value
through a shared reference, and can do so safely as long as they use atomic operations. Any
use of the returned raw pointer requires an unsafe
block and still has to uphold the same
restriction: operations on it must be atomic.
§Examples
use std::sync::atomic::AtomicUsize;
extern "C" {
fn my_atomic_op(arg: *mut usize);
}
let atomic = AtomicUsize::new(1);
// SAFETY: Safe as long as `my_atomic_op` is atomic.
unsafe {
my_atomic_op(atomic.as_ptr());
}
Trait Implementations§
source§impl BitStore for AtomicUsize
impl BitStore for AtomicUsize
§type Mem = usize
type Mem = usize
BitSlice
. It
is always one of the unsigned integer fundamentals.§type Access = AtomicUsize
type Access = AtomicUsize
Self::Mem
value between the processor and the memory system. Read more§type Alias = AtomicUsize
type Alias = AtomicUsize
BitStore
implementor that is known to be alias-safe. It is
used when a BitSlice
introduces multiple handles that view the same
memory location, and at least one of them has write capabilities to it.
It must have the same underlying memory type, and can only change access
patterns or public-facing usage.§type Unalias = AtomicUsize
type Unalias = AtomicUsize
::Alias
. It is used when a BitSlice
removes the
conditions that required a T -> T::Alias
transition.source§const ZERO: AtomicUsize = _
const ZERO: AtomicUsize = _
source§fn new(value: <AtomicUsize as BitStore>::Mem) -> AtomicUsize
fn new(value: <AtomicUsize as BitStore>::Mem) -> AtomicUsize
BitStore
type.source§fn load_value(&self) -> <AtomicUsize as BitStore>::Mem
fn load_value(&self) -> <AtomicUsize as BitStore>::Mem
::Access
rules. This may be called when the value is aliased by a write-capable
reference.source§fn store_value(&mut self, value: <AtomicUsize as BitStore>::Mem)
fn store_value(&mut self, value: <AtomicUsize as BitStore>::Mem)
::Access
constraints.source§const ALIGNED_TO_SIZE: [(); 1] = _
const ALIGNED_TO_SIZE: [(); 1] = _
1.3.0 · source§impl Debug for AtomicUsize
impl Debug for AtomicUsize
1.0.0 · source§impl Default for AtomicUsize
impl Default for AtomicUsize
source§fn default() -> AtomicUsize
fn default() -> AtomicUsize
source§impl<'de> Deserialize<'de> for AtomicUsize
impl<'de> Deserialize<'de> for AtomicUsize
source§fn deserialize<D>(
deserializer: D,
) -> Result<AtomicUsize, <D as Deserializer<'de>>::Error>where
D: Deserializer<'de>,
fn deserialize<D>(
deserializer: D,
) -> Result<AtomicUsize, <D as Deserializer<'de>>::Error>where
D: Deserializer<'de>,
1.23.0 · source§impl From<usize> for AtomicUsize
impl From<usize> for AtomicUsize
source§fn from(v: usize) -> AtomicUsize
fn from(v: usize) -> AtomicUsize
Converts an usize
into an AtomicUsize
.
source§impl Radium for AtomicUsize
impl Radium for AtomicUsize
type Item = usize
source§fn new(value: usize) -> AtomicUsize
fn new(value: usize) -> AtomicUsize
source§fn get_mut(&mut self) -> &mut usize
fn get_mut(&mut self) -> &mut usize
source§fn into_inner(self) -> usize
fn into_inner(self) -> usize
source§fn swap(&self, value: usize, order: Ordering) -> usize
fn swap(&self, value: usize, order: Ordering) -> usize
source§fn compare_and_swap(&self, current: usize, new: usize, order: Ordering) -> usize
fn compare_and_swap(&self, current: usize, new: usize, order: Ordering) -> usize
compare_exchange
or compare_exchange_weak
insteadcurrent
value. Read moresource§fn compare_exchange(
&self,
current: usize,
new: usize,
success: Ordering,
failure: Ordering,
) -> Result<usize, usize>
fn compare_exchange( &self, current: usize, new: usize, success: Ordering, failure: Ordering, ) -> Result<usize, usize>
current
value. Read moresource§fn compare_exchange_weak(
&self,
current: usize,
new: usize,
success: Ordering,
failure: Ordering,
) -> Result<usize, usize>
fn compare_exchange_weak( &self, current: usize, new: usize, success: Ordering, failure: Ordering, ) -> Result<usize, usize>
current
value. Read moresource§fn fetch_update<F>(
&self,
set_order: Ordering,
fetch_order: Ordering,
f: F,
) -> Result<usize, usize>
fn fetch_update<F>( &self, set_order: Ordering, fetch_order: Ordering, f: F, ) -> Result<usize, usize>
source§fn fetch_and(&self, value: usize, order: Ordering) -> usize
fn fetch_and(&self, value: usize, order: Ordering) -> usize
value
, and stores the result in self
. Read moresource§fn fetch_nand(&self, value: usize, order: Ordering) -> usize
fn fetch_nand(&self, value: usize, order: Ordering) -> usize
value
, and stores the result in self
. Read moresource§fn fetch_or(&self, value: usize, order: Ordering) -> usize
fn fetch_or(&self, value: usize, order: Ordering) -> usize
value
, and stores the result in self
. Read moresource§fn fetch_xor(&self, value: usize, order: Ordering) -> usize
fn fetch_xor(&self, value: usize, order: Ordering) -> usize
value
, and stores the result in self
. Read moresource§impl Serialize for AtomicUsize
impl Serialize for AtomicUsize
source§fn serialize<S>(
&self,
serializer: S,
) -> Result<<S as Serializer>::Ok, <S as Serializer>::Error>where
S: Serializer,
fn serialize<S>(
&self,
serializer: S,
) -> Result<<S as Serializer>::Ok, <S as Serializer>::Error>where
S: Serializer,
impl RefUnwindSafe for AtomicUsize
impl Sync for AtomicUsize
Auto Trait Implementations§
impl !Freeze for AtomicUsize
impl Send for AtomicUsize
impl Unpin for AtomicUsize
impl UnwindSafe for AtomicUsize
Blanket Implementations§
source§impl<A> BitAccess for A
impl<A> BitAccess for A
source§fn clear_bits(&self, mask: BitMask<Self::Item>) -> Self::Item
fn clear_bits(&self, mask: BitMask<Self::Item>) -> Self::Item
0
. Read moresource§fn set_bits(&self, mask: BitMask<Self::Item>) -> Self::Item
fn set_bits(&self, mask: BitMask<Self::Item>) -> Self::Item
1
. Read moresource§fn invert_bits(&self, mask: BitMask<Self::Item>) -> Self::Item
fn invert_bits(&self, mask: BitMask<Self::Item>) -> Self::Item
source§impl<T> BitView for Twhere
T: BitStore,
impl<T> BitView for Twhere
T: BitStore,
source§fn view_bits<O>(&self) -> &BitSlice<T, O> ⓘwhere
O: BitOrder,
fn view_bits<O>(&self) -> &BitSlice<T, O> ⓘwhere
O: BitOrder,
source§fn try_view_bits<O>(&self) -> Result<&BitSlice<T, O>, BitSpanError<T>>where
O: BitOrder,
fn try_view_bits<O>(&self) -> Result<&BitSlice<T, O>, BitSpanError<T>>where
O: BitOrder,
source§fn view_bits_mut<O>(&mut self) -> &mut BitSlice<T, O> ⓘwhere
O: BitOrder,
fn view_bits_mut<O>(&mut self) -> &mut BitSlice<T, O> ⓘwhere
O: BitOrder,
source§fn try_view_bits_mut<O>(
&mut self,
) -> Result<&mut BitSlice<T, O>, BitSpanError<T>>where
O: BitOrder,
fn try_view_bits_mut<O>(
&mut self,
) -> Result<&mut BitSlice<T, O>, BitSpanError<T>>where
O: BitOrder,
source§impl<T> BitViewSized for Twhere
T: BitStore,
impl<T> BitViewSized for Twhere
T: BitStore,
source§fn as_raw_slice(&self) -> &[<T as BitView>::Store]
fn as_raw_slice(&self) -> &[<T as BitView>::Store]
source§fn as_raw_mut_slice(&mut self) -> &mut [<T as BitView>::Store]
fn as_raw_mut_slice(&mut self) -> &mut [<T as BitView>::Store]
source§fn into_bitarray<O>(self) -> BitArray<Self, O>where
O: BitOrder,
fn into_bitarray<O>(self) -> BitArray<Self, O>where
O: BitOrder,
self
in a BitArray
.source§impl<T> BorrowMut<T> for Twhere
T: ?Sized,
impl<T> BorrowMut<T> for Twhere
T: ?Sized,
source§fn borrow_mut(&mut self) -> &mut T
fn borrow_mut(&mut self) -> &mut T
source§impl<T> FmtForward for T
impl<T> FmtForward for T
source§fn fmt_binary(self) -> FmtBinary<Self>where
Self: Binary,
fn fmt_binary(self) -> FmtBinary<Self>where
Self: Binary,
self
to use its Binary
implementation when Debug
-formatted.source§fn fmt_display(self) -> FmtDisplay<Self>where
Self: Display,
fn fmt_display(self) -> FmtDisplay<Self>where
Self: Display,
self
to use its Display
implementation when
Debug
-formatted.source§fn fmt_lower_exp(self) -> FmtLowerExp<Self>where
Self: LowerExp,
fn fmt_lower_exp(self) -> FmtLowerExp<Self>where
Self: LowerExp,
self
to use its LowerExp
implementation when
Debug
-formatted.source§fn fmt_lower_hex(self) -> FmtLowerHex<Self>where
Self: LowerHex,
fn fmt_lower_hex(self) -> FmtLowerHex<Self>where
Self: LowerHex,
self
to use its LowerHex
implementation when
Debug
-formatted.source§fn fmt_octal(self) -> FmtOctal<Self>where
Self: Octal,
fn fmt_octal(self) -> FmtOctal<Self>where
Self: Octal,
self
to use its Octal
implementation when Debug
-formatted.source§fn fmt_pointer(self) -> FmtPointer<Self>where
Self: Pointer,
fn fmt_pointer(self) -> FmtPointer<Self>where
Self: Pointer,
self
to use its Pointer
implementation when
Debug
-formatted.source§fn fmt_upper_exp(self) -> FmtUpperExp<Self>where
Self: UpperExp,
fn fmt_upper_exp(self) -> FmtUpperExp<Self>where
Self: UpperExp,
self
to use its UpperExp
implementation when
Debug
-formatted.source§fn fmt_upper_hex(self) -> FmtUpperHex<Self>where
Self: UpperHex,
fn fmt_upper_hex(self) -> FmtUpperHex<Self>where
Self: UpperHex,
self
to use its UpperHex
implementation when
Debug
-formatted.source§impl<T> Pipe for Twhere
T: ?Sized,
impl<T> Pipe for Twhere
T: ?Sized,
source§fn pipe<R>(self, func: impl FnOnce(Self) -> R) -> Rwhere
Self: Sized,
fn pipe<R>(self, func: impl FnOnce(Self) -> R) -> Rwhere
Self: Sized,
source§fn pipe_ref<'a, R>(&'a self, func: impl FnOnce(&'a Self) -> R) -> Rwhere
R: 'a,
fn pipe_ref<'a, R>(&'a self, func: impl FnOnce(&'a Self) -> R) -> Rwhere
R: 'a,
self
and passes that borrow into the pipe function. Read moresource§fn pipe_ref_mut<'a, R>(&'a mut self, func: impl FnOnce(&'a mut Self) -> R) -> Rwhere
R: 'a,
fn pipe_ref_mut<'a, R>(&'a mut self, func: impl FnOnce(&'a mut Self) -> R) -> Rwhere
R: 'a,
self
and passes that borrow into the pipe function. Read moresource§fn pipe_borrow<'a, B, R>(&'a self, func: impl FnOnce(&'a B) -> R) -> R
fn pipe_borrow<'a, B, R>(&'a self, func: impl FnOnce(&'a B) -> R) -> R
source§fn pipe_borrow_mut<'a, B, R>(
&'a mut self,
func: impl FnOnce(&'a mut B) -> R,
) -> R
fn pipe_borrow_mut<'a, B, R>( &'a mut self, func: impl FnOnce(&'a mut B) -> R, ) -> R
source§fn pipe_as_ref<'a, U, R>(&'a self, func: impl FnOnce(&'a U) -> R) -> R
fn pipe_as_ref<'a, U, R>(&'a self, func: impl FnOnce(&'a U) -> R) -> R
self
, then passes self.as_ref()
into the pipe function.source§fn pipe_as_mut<'a, U, R>(&'a mut self, func: impl FnOnce(&'a mut U) -> R) -> R
fn pipe_as_mut<'a, U, R>(&'a mut self, func: impl FnOnce(&'a mut U) -> R) -> R
self
, then passes self.as_mut()
into the pipe
function.source§fn pipe_deref<'a, T, R>(&'a self, func: impl FnOnce(&'a T) -> R) -> R
fn pipe_deref<'a, T, R>(&'a self, func: impl FnOnce(&'a T) -> R) -> R
self
, then passes self.deref()
into the pipe function.source§impl<T> Tap for T
impl<T> Tap for T
source§fn tap_borrow<B>(self, func: impl FnOnce(&B)) -> Self
fn tap_borrow<B>(self, func: impl FnOnce(&B)) -> Self
Borrow<B>
of a value. Read moresource§fn tap_borrow_mut<B>(self, func: impl FnOnce(&mut B)) -> Self
fn tap_borrow_mut<B>(self, func: impl FnOnce(&mut B)) -> Self
BorrowMut<B>
of a value. Read moresource§fn tap_ref<R>(self, func: impl FnOnce(&R)) -> Self
fn tap_ref<R>(self, func: impl FnOnce(&R)) -> Self
AsRef<R>
view of a value. Read moresource§fn tap_ref_mut<R>(self, func: impl FnOnce(&mut R)) -> Self
fn tap_ref_mut<R>(self, func: impl FnOnce(&mut R)) -> Self
AsMut<R>
view of a value. Read moresource§fn tap_deref<T>(self, func: impl FnOnce(&T)) -> Self
fn tap_deref<T>(self, func: impl FnOnce(&T)) -> Self
Deref::Target
of a value. Read moresource§fn tap_deref_mut<T>(self, func: impl FnOnce(&mut T)) -> Self
fn tap_deref_mut<T>(self, func: impl FnOnce(&mut T)) -> Self
Deref::Target
of a value. Read moresource§fn tap_dbg(self, func: impl FnOnce(&Self)) -> Self
fn tap_dbg(self, func: impl FnOnce(&Self)) -> Self
.tap()
only in debug builds, and is erased in release builds.source§fn tap_mut_dbg(self, func: impl FnOnce(&mut Self)) -> Self
fn tap_mut_dbg(self, func: impl FnOnce(&mut Self)) -> Self
.tap_mut()
only in debug builds, and is erased in release
builds.source§fn tap_borrow_dbg<B>(self, func: impl FnOnce(&B)) -> Self
fn tap_borrow_dbg<B>(self, func: impl FnOnce(&B)) -> Self
.tap_borrow()
only in debug builds, and is erased in release
builds.source§fn tap_borrow_mut_dbg<B>(self, func: impl FnOnce(&mut B)) -> Self
fn tap_borrow_mut_dbg<B>(self, func: impl FnOnce(&mut B)) -> Self
.tap_borrow_mut()
only in debug builds, and is erased in release
builds.source§fn tap_ref_dbg<R>(self, func: impl FnOnce(&R)) -> Self
fn tap_ref_dbg<R>(self, func: impl FnOnce(&R)) -> Self
.tap_ref()
only in debug builds, and is erased in release
builds.source§fn tap_ref_mut_dbg<R>(self, func: impl FnOnce(&mut R)) -> Self
fn tap_ref_mut_dbg<R>(self, func: impl FnOnce(&mut R)) -> Self
.tap_ref_mut()
only in debug builds, and is erased in release
builds.source§fn tap_deref_dbg<T>(self, func: impl FnOnce(&T)) -> Self
fn tap_deref_dbg<T>(self, func: impl FnOnce(&T)) -> Self
.tap_deref()
only in debug builds, and is erased in release
builds.