correct timing stack reordering in am_now_working_on and add separate timing category for or_to_all #952
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Two improvements to the timing statistics feature:
am_now_working_on
by reversing the order (going from top to bottom) in which the timing stackwas_working_on
was being reshuffled upon the addition of a new entry at the bottom (i.e. 0th element)MpiTime
) for two separate instances ofor_to_all
; one of these (inphase_material
) was causing synchronization during the time stepping which was mixing up the results.With these changes, the timing statistics for two examples involving a tall, skinny, 2d cell with a
FluxRegion
and a Lorentzianmaterial
are now as expected: when run with 3 processors/chunks, the "waiting" time for two of the three chunks (which do not contain the additional expensive pixels) are contained in thecommunicating:
bin rather than intime stepping:
as was reported previously.tall, skinny cell with
FluxRegion
output for 3 processors/chunks
tall, skinny cell with Lorentzian
material
output for 3 processors/chunks