diff options
| author | Douglas Rumbaugh <dbr4@psu.edu> | 2023-11-13 11:44:09 -0500 |
|---|---|---|
| committer | Douglas Rumbaugh <dbr4@psu.edu> | 2023-11-13 11:44:09 -0500 |
| commit | 90bb0614fc1d8f1a185a778e31aaf9027c01aeb8 (patch) | |
| tree | 22a865978e398cef2fdcef763ab893c2da769a4a /include/framework/structure/MutableBuffer.h | |
| parent | 83486744600e8be338c75c2e3d2339452a392a9d (diff) | |
| download | dynamic-extension-90bb0614fc1d8f1a185a778e31aaf9027c01aeb8.tar.gz | |
Tombstone Compaction: re-enabled tombstone compaction
Currently, proactive buffer tombstone compaction is disabled by forcing
the buffer tombstone capacity to match its record capacity. It isn't
clear how to best handle proactive buffer compactions in an environment
where new buffers are spawned anyway.
Diffstat (limited to 'include/framework/structure/MutableBuffer.h')
| -rw-r--r-- | include/framework/structure/MutableBuffer.h | 2 |
1 files changed, 1 insertions, 1 deletions
diff --git a/include/framework/structure/MutableBuffer.h b/include/framework/structure/MutableBuffer.h index 671824f..8b17091 100644 --- a/include/framework/structure/MutableBuffer.h +++ b/include/framework/structure/MutableBuffer.h @@ -32,7 +32,7 @@ template <RecordInterface R> class MutableBuffer { public: MutableBuffer(size_t capacity, size_t max_tombstone_cap) - : m_cap(capacity), m_tombstone_cap(max_tombstone_cap), m_reccnt(0) + : m_cap(capacity), m_tombstone_cap(capacity), m_reccnt(0) , m_tombstonecnt(0), m_weight(0), m_max_weight(0), m_tail(0) { m_data = (Wrapped<R>*) psudb::sf_aligned_alloc(CACHELINE_SIZE, capacity*sizeof(Wrapped<R>)); m_merge_data = (Wrapped<R>*) psudb::sf_aligned_alloc(CACHELINE_SIZE, capacity*sizeof(Wrapped<R>)); |