[mvapich-discuss] job metrics

Walid walid.shaari at gmail.com
Tue Apr 29 21:55:01 EDT 2014


Hari, Jennifer,

Thanks for your feedback, i am representing the system side in my
organisation, and application support would like to have these metrics as a
regular feature, part of the job output. what i understand from both of you
and a colleague of mine that you usually do profiling as part of a
performance exercise, or to understand more about the job characteristics,
we do use Infiniband and ethernet as interconnects, and some of our jobs do
extend large number of cores up in 100s, and sometime a couple of 1000
cores. would not that be collecting traces from every node and writing it
causes additional overhead, does the collection and saving of metrics
scale? apart from profilers do you do anything different to monitor job
trends, and correlate behaviour, or capacity planing?

kind regards

Walid


On 30 April 2014 00:33, Hari Subramoni <subramoni.1 at osu.edu> wrote:

> Hello Walid,
>
> There are several profilers like mpiP, IPM, TAU and HPCToolkit that allows
> folks to characterize the job and workload in terms of number of / time
> spent in MPI calls with relatively low overheads. The amount of overhead
> cause and the level and kind of detail provided differs from profiler to
> profiler.
>
> Thx,
> Hari.
>
>
> On Tue, Apr 29, 2014 at 5:08 AM, Walid <walid.shaari at gmail.com> wrote:
>
>> Hi,
>>
>> What would be the best way to characterize the job and workload it
>> processed in terms of mpi calls and communications, how is easy to profile
>> a job run and how much overhead does the profiling cause?
>>
>> any one doing extensive parallel jobs profiling, capacity planing that
>> would like to share their thoughts?
>>
>> kind regards
>>
>> Walid
>>
>> _______________________________________________
>> mvapich-discuss mailing list
>> mvapich-discuss at cse.ohio-state.edu
>> http://mailman.cse.ohio-state.edu/mailman/listinfo/mvapich-discuss
>>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://mailman.cse.ohio-state.edu/pipermail/mvapich-discuss/attachments/20140430/848addc5/attachment-0001.html>


More information about the mvapich-discuss mailing list