+void Segment::get_raw_samples(uint64_t start, uint64_t count,
+ uint8_t* dest) const
+{
+ assert(start < sample_count_);
+ assert(start + count <= sample_count_);
+ assert(count > 0);
+ assert(dest != nullptr);
+
+ lock_guard<recursive_mutex> lock(mutex_);
+
+ uint8_t* dest_ptr = dest;
+
+ uint64_t chunk_num = (start * unit_size_) / chunk_size_;
+ uint64_t chunk_offs = (start * unit_size_) % chunk_size_;
+
+ while (count > 0) {
+ const uint8_t* chunk = data_chunks_[chunk_num];
+
+ uint64_t copy_size = min(count * unit_size_,
+ chunk_size_ - chunk_offs);
+
+ memcpy(dest_ptr, chunk + chunk_offs, copy_size);
+
+ dest_ptr += copy_size;
+ count -= (copy_size / unit_size_);
+
+ chunk_num++;
+ chunk_offs = 0;
+ }
+}
+
+SegmentDataIterator* Segment::begin_sample_iteration(uint64_t start)
+{
+ SegmentDataIterator* it = new SegmentDataIterator;
+
+ assert(start < sample_count_);
+
+ iterator_count_++;
+
+ it->sample_index = start;
+ it->chunk_num = (start * unit_size_) / chunk_size_;
+ it->chunk_offs = (start * unit_size_) % chunk_size_;
+ it->chunk = data_chunks_[it->chunk_num];
+
+ return it;
+}
+
+void Segment::continue_sample_iteration(SegmentDataIterator* it, uint64_t increase)
+{
+ it->sample_index += increase;
+ it->chunk_offs += (increase * unit_size_);
+
+ if (it->chunk_offs > (chunk_size_ - 1)) {
+ it->chunk_num++;
+ it->chunk_offs -= chunk_size_;
+ it->chunk = data_chunks_[it->chunk_num];
+ }
+}
+
+void Segment::end_sample_iteration(SegmentDataIterator* it)
+{
+ delete it;
+
+ iterator_count_--;
+
+ if ((iterator_count_ == 0) && mem_optimization_requested_) {
+ mem_optimization_requested_ = false;
+ free_unused_memory();
+ }
+}
+
+uint8_t* Segment::get_iterator_value(SegmentDataIterator* it)
+{
+ assert(it->sample_index <= (sample_count_ - 1));
+
+ return (it->chunk + it->chunk_offs);
+}
+
+uint64_t Segment::get_iterator_valid_length(SegmentDataIterator* it)
+{
+ assert(it->sample_index <= (sample_count_ - 1));
+
+ return ((chunk_size_ - it->chunk_offs) / unit_size_);
+}
+