diff --git a/src/chunks.rs b/src/chunks.rs index dad49ba..6f6490d 100644 --- a/src/chunks.rs +++ b/src/chunks.rs @@ -237,7 +237,7 @@ impl Producer { /// For a safe alternative that provides mutable slices of [`Default`]-initialized slots, /// see [`Producer::write_chunk()`]. pub fn write_chunk_uninit(&mut self, n: usize) -> Result, ChunkError> { - let tail = self.cached_tail.get(); + let tail = self.cached_tail; // Check if the queue has *possibly* not enough slots. if self.buffer.capacity - self.buffer.distance(self.cached_head.get(), tail) < n { @@ -286,7 +286,7 @@ impl Consumer { /// /// See the documentation of the [`chunks`](crate::chunks#examples) module. pub fn read_chunk(&mut self, n: usize) -> Result, ChunkError> { - let head = self.cached_head.get(); + let head = self.cached_head; // Check if the queue has *possibly* not enough slots. if self.buffer.distance(head, self.cached_tail.get()) < n { @@ -441,7 +441,7 @@ pub struct WriteChunkUninit<'a, T> { first_len: usize, second_ptr: *mut T, second_len: usize, - producer: &'a Producer, + producer: &'a mut Producer, } impl WriteChunkUninit<'_, T> { @@ -494,9 +494,9 @@ impl WriteChunkUninit<'_, T> { unsafe fn commit_unchecked(self, n: usize) -> usize { let p = self.producer; - let tail = p.buffer.increment(p.cached_tail.get(), n); + let tail = p.buffer.increment(p.cached_tail, n); p.buffer.tail.store(tail, Ordering::Release); - p.cached_tail.set(tail); + p.cached_tail = tail; n } @@ -707,9 +707,9 @@ impl ReadChunk<'_, T> { self.second_ptr.add(i).drop_in_place(); } let c = self.consumer; - let head = c.buffer.increment(c.cached_head.get(), n); + let head = c.buffer.increment(c.cached_head, n); c.buffer.head.store(head, Ordering::Release); - c.cached_head.set(head); + c.cached_head = head; n } @@ -760,10 +760,10 @@ impl<'a, T> Drop for ReadChunkIntoIter<'a, T> { /// /// Non-iterated items remain in the ring buffer and are *not* dropped. fn drop(&mut self) { - let c = &self.chunk.consumer; - let head = c.buffer.increment(c.cached_head.get(), self.iterated); + let c = &mut self.chunk.consumer; + let head = c.buffer.increment(c.cached_head, self.iterated); c.buffer.head.store(head, Ordering::Release); - c.cached_head.set(head); + c.cached_head = head; } } diff --git a/src/lib.rs b/src/lib.rs index 037a4ee..d8e527b 100644 --- a/src/lib.rs +++ b/src/lib.rs @@ -127,11 +127,11 @@ impl RingBuffer { let p = Producer { buffer: buffer.clone(), cached_head: Cell::new(0), - cached_tail: Cell::new(0), + cached_tail: 0, }; let c = Consumer { buffer, - cached_head: Cell::new(0), + cached_head: 0, cached_tail: Cell::new(0), }; (p, c) @@ -285,7 +285,7 @@ pub struct Producer { /// A copy of `buffer.tail` for quick access. /// /// This value is always in sync with `buffer.tail`. - cached_tail: Cell, + cached_tail: usize, } unsafe impl Send for Producer {} @@ -317,7 +317,7 @@ impl Producer { } let tail = self.buffer.increment1(tail); self.buffer.tail.store(tail, Ordering::Release); - self.cached_tail.set(tail); + self.cached_tail = tail; Ok(()) } else { Err(PushError::Full(value)) @@ -345,7 +345,7 @@ impl Producer { pub fn slots(&self) -> usize { let head = self.buffer.head.load(Ordering::Acquire); self.cached_head.set(head); - self.buffer.capacity - self.buffer.distance(head, self.cached_tail.get()) + self.buffer.capacity - self.buffer.distance(head, self.cached_tail) } /// Returns `true` if there are currently no slots available for writing. @@ -440,7 +440,7 @@ impl Producer { /// This is a strict subset of the functionality implemented in `write_chunk_uninit()`. /// For performance, this special case is immplemented separately. fn next_tail(&self) -> Option { - let tail = self.cached_tail.get(); + let tail = self.cached_tail; // Check if the queue is *possibly* full. if self.buffer.distance(self.cached_head.get(), tail) == self.buffer.capacity { @@ -485,7 +485,7 @@ pub struct Consumer { /// A copy of `buffer.head` for quick access. /// /// This value is always in sync with `buffer.head`. - cached_head: Cell, + cached_head: usize, /// A copy of `buffer.tail` for quick access. /// @@ -530,7 +530,7 @@ impl Consumer { let value = unsafe { self.buffer.slot_ptr(head).read() }; let head = self.buffer.increment1(head); self.buffer.head.store(head, Ordering::Release); - self.cached_head.set(head); + self.cached_head = head; Ok(value) } else { Err(PopError::Empty) @@ -584,7 +584,7 @@ impl Consumer { pub fn slots(&self) -> usize { let tail = self.buffer.tail.load(Ordering::Acquire); self.cached_tail.set(tail); - self.buffer.distance(self.cached_head.get(), tail) + self.buffer.distance(self.cached_head, tail) } /// Returns `true` if there are currently no slots available for reading. @@ -678,7 +678,7 @@ impl Consumer { /// This is a strict subset of the functionality implemented in `read_chunk()`. /// For performance, this special case is immplemented separately. fn next_head(&self) -> Option { - let head = self.cached_head.get(); + let head = self.cached_head; // Check if the queue is *possibly* empty. if head == self.cached_tail.get() {