Drain TriggerRecords from the DataWriter input queue at end-run#476
Merged
bieryAtFnal merged 3 commits intodevelopfrom Feb 11, 2026
Merged
Drain TriggerRecords from the DataWriter input queue at end-run#476bieryAtFnal merged 3 commits intodevelopfrom
bieryAtFnal merged 3 commits intodevelopfrom
Conversation
added 3 commits
January 21, 2026 08:48
…from the input queue before stopping.
…k to use the new receiver data_pending() method.
…rocessAllEventsAtStop
This was referenced Feb 9, 2026
eflumerf
approved these changes
Feb 10, 2026
Member
eflumerf
left a comment
There was a problem hiding this comment.
Following Kurt's excellent test procedure, I saw the expected failure and fix.
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
Description
Back in December, Michal reported that DFO errors were seen in data-taking runs at NP02 at end-run ("stop") time. The errors were along the lines of the following:
We believe that these errors were the result of trying to write more data than the configured number of storage disks could handle. (The system was clogged, and the DFO didn't wait very long for the backup to clear - just 100 msec.)
A problem that was noticed when we investigated these messages is that TriggerRecords can get lost when this happens.
The changes in this PR attempt to avoid that problem by reading all TriggerRecords in the DataWriter input queue before stopping at end-run time.
This change is correlated with the following ones:
Those changes need to be merged before, or at the same time, as this one.
Here are suggested instructions for demonstrating the problem and validating the fix:
For reference, there are a few notes on this topic in the agenda of the 21-Jan-2026 Dataflow WG meeting.
Type of change
Testing checklist
dbt-build --unittest)pytest -s minimal_system_quick_test.py)daqsystemtest_integtest_bundle.sh)Further checks