29 if (!bufPage.fPage.IsNull()) {
46 "wall clock time spent in critical sections"),
48 "timeCpuCriticalSection",
"ns",
"CPU time spent in critical section")});
72 f.SetOnDiskId(fNFields);
75 for (
auto *
f : fields) {
77 for (
auto &descendant : *
f) {
78 connectField(descendant);
81 fBufferedColumns.resize(fNColumns);
86 return fInnerSink->GetDescriptor();
93 fInnerModel = model.
Clone();
94 fInnerSink->Init(*fInnerModel);
104 auto cloneAddField = [&](
const RFieldBase *field) {
105 auto cloned = field->Clone(field->GetFieldName());
107 fInnerModel->AddField(std::move(cloned));
110 auto cloneAddProjectedField = [&](
RFieldBase *field) {
111 auto cloned = field->
Clone(field->GetFieldName());
116 auto targetIt = cloned->
begin();
117 for (
auto &
f : *field)
118 fieldMap[&(*targetIt++)] = projectedFields.GetSourceField(&
f);
123 fInnerModel->Unfreeze();
125 std::back_inserter(innerChangeset.fAddedFields), cloneAddField);
127 std::back_inserter(innerChangeset.fAddedProjectedFields), cloneAddProjectedField);
128 fInnerModel->Freeze();
129 fInnerSink->UpdateSchema(innerChangeset, firstEntry);
141 auto &zipItem = fBufferedColumns.at(colId).BufferPage(columnHandle);
142 zipItem.AllocateSealedPageBuf(page.
GetNBytes());
144 auto &sealedPage = fBufferedColumns.at(colId).RegisterSealedPage();
146 if (!fTaskScheduler) {
149 SealPage(page, element, GetWriteOptions().GetCompression(), zipItem.fBuf.get(),
false);
150 zipItem.fSealedPage = &sealedPage;
155 zipItem.fPage = ReservePage(columnHandle, page.
GetNElements());
160 fCounters->fParallelZip.SetValue(1);
163 fTaskScheduler->AddTask([
this, &zipItem, &sealedPage, &element] {
164 sealedPage = SealPage(zipItem.fPage, element, GetWriteOptions().GetCompression(), zipItem.fBuf.get());
165 zipItem.fSealedPage = &sealedPage;
184 std::vector<RSealedPageGroup> toCommit;
185 toCommit.reserve(fBufferedColumns.size());
186 for (
auto &bufColumn : fBufferedColumns) {
187 R__ASSERT(bufColumn.HasSealedPagesOnly());
188 const auto &sealedPages = bufColumn.GetSealedPages();
189 toCommit.emplace_back(bufColumn.GetHandle().fPhysicalId, sealedPages.cbegin(), sealedPages.cend());
192 std::uint64_t nbytes;
196 fInnerSink->CommitSealedPageV(toCommit);
198 nbytes = fInnerSink->CommitCluster(nNewEntries);
201 for (
auto &bufColumn : fBufferedColumns)
202 bufColumn.DropBufferedPages();
210 fInnerSink->CommitClusterGroup();
217 fInnerSink->CommitDataset();
223 return fInnerSink->ReservePage(columnHandle, nElements);
228 fInnerSink->ReleasePage(page);
#define R__FAIL(msg)
Short-hand to return an RResult<T> in an error state; the RError is implicitly converted into RResult...
winID h TVirtualViewer3D TVirtualGLPainter p
A collection of Counter objects with a name, a unit, and a description.
void ObserveMetrics(RNTupleMetrics &observee)
CounterPtrT MakeCounter(const std::string &name, Args &&... args)
A non thread-safe integral performance counter.
An either thread-safe or non thread safe counter for CPU ticks.
Record wall time and CPU time between construction and destruction.
RColumnElementBase * GetElement() const
RPageSink * GetPageSink() const
RPageStorage::SealedPageSequence_t fSealedPages
Pages that have been already sealed by a concurrent task.
std::deque< RPageZipItem > fBufferedPages
Using a deque guarantees that element iterators are never invalidated by appends to the end of the it...
RPageStorage::ColumnHandle_t fCol
void CommitDataset() final
Finalize the current cluster and the entrire data set.
std::uint64_t CommitCluster(NTupleSize_t nNewEntries) final
Finalize the current cluster and create a new one for the following data.
std::unique_ptr< RPageSink > fInnerSink
The inner sink, responsible for actually performing I/O.
RPage ReservePage(ColumnHandle_t columnHandle, std::size_t nElements) final
Get a new, empty page for the given column that can be filled with up to nElements.
void CommitSealedPageV(std::span< RPageStorage::RSealedPageGroup > ranges) final
Write a vector of preprocessed pages to storage. The corresponding columns must have been added befor...
void ReleasePage(RPage &page) final
Every page store needs to be able to free pages it handed out.
void UpdateSchema(const RNTupleModelChangeset &changeset, NTupleSize_t firstEntry) final
Incorporate incremental changes to the model into the ntuple descriptor.
RPageSinkBuf(std::unique_ptr< RPageSink > inner)
const RNTupleDescriptor & GetDescriptor() const final
Return the RNTupleDescriptor being constructed.
std::unique_ptr< RCounters > fCounters
ColumnHandle_t AddColumn(DescriptorId_t fieldId, const RColumn &column) final
Register a new column.
void InitImpl(RNTupleModel &model) final
void CommitPage(ColumnHandle_t columnHandle, const RPage &page) final
Write a page to the storage. The column must have been added before.
void ConnectFields(const std::vector< RFieldBase * > &fields, NTupleSize_t firstEntry)
void CommitSealedPage(DescriptorId_t physicalColumnId, const RSealedPage &sealedPage) final
Write a preprocessed page to storage. The column must have been added before.
void CommitClusterGroup() final
Write out the page locations (page list envelope) for all the committed clusters since the last call ...
An RAII wrapper used to synchronize a page sink. See GetSinkGuard().
Abstract interface to write data into an ntuple.
const RNTupleWriteOptions & GetWriteOptions() const
Returns the sink's write options.
const std::string & GetNTupleName() const
Returns the NTuple name.
virtual void ReleasePage(RPage &page)=0
Every page store needs to be able to free pages it handed out.
Detail::RNTupleMetrics fMetrics
A page is a slice of a column that is mapped into memory.
std::uint32_t GetNBytes() const
The space taken by column elements in the buffer.
std::uint32_t GetNElements() const
Base class for all ROOT issued exceptions.
A field translates read and write calls from/to underlying columns to/from tree values.
std::vector< RFieldBase * > GetSubFields()
The on-storage meta-data of an ntuple.
Projected fields are fields whose columns are reused from existing fields.
std::unordered_map< const RFieldBase *, const RFieldBase * > FieldMap_t
The map keys are the projected target fields, the map values are the backing source fields Note that ...
const RFieldBase * GetSourceField(const RFieldBase *target) const
The RNTupleModel encapulates the schema of an ntuple.
std::unique_ptr< RNTupleModel > Clone() const
const RProjectedFields & GetProjectedFields() const
RFieldZero & GetFieldZero()
Non-const access to the root field is used to commit clusters during writing and to set the on-disk f...
virtual TObject * Clone(const char *newname="") const
Make a clone of an object using the Streamer facility.
void CallConnectPageSinkOnField(RFieldBase &, RPageSink &, NTupleSize_t firstEntry=0)
void Add(RHist< DIMENSIONS, PRECISION, STAT_TO... > &to, const RHist< DIMENSIONS, PRECISION, STAT_FROM... > &from)
Add two histograms.
std::uint64_t NTupleSize_t
Integer type long enough to hold the maximum number of entries in a column.
std::uint64_t DescriptorId_t
Distriniguishes elements of the same type within a descriptor, e.g. different fields.
The incremental changes to a RNTupleModel
std::vector< RFieldBase * > fAddedProjectedFields
Points to the projected fields in fModel that were added as part of an updater transaction.
std::vector< RFieldBase * > fAddedFields
Points to the fields in fModel that were added as part of an updater transaction.
I/O performance counters that get registered in fMetrics.
DescriptorId_t fPhysicalId
A sealed page contains the bytes of a page as written to storage (packed & compressed).