/[escript]/trunk/escript/py_src/benchmark.py
ViewVC logotype

Annotation of /trunk/escript/py_src/benchmark.py

Parent Directory Parent Directory | Revision Log Revision Log


Revision 386 - (hide annotations)
Tue Dec 20 00:31:52 2005 UTC (13 years, 9 months ago) by gross
File MIME type: text/x-python
File size: 13002 byte(s)
the number of problems to run is now the less or equal then len(scale) and less or equal number of problems. This allows easier control
1 gross 379 filter# $Id:$
2 gross 364
3     #
4     # COPYRIGHT ACcESS 2004 - All Rights Reserved
5     #
6     # This software is the property of ACcESS. No part of this code
7     # may be copied in any form or by any means without the expressed written
8     # consent of ACcESS. Copying, use or modification of this software
9     # by any unauthorised person is illegal unless that
10     # person has a software license agreement with ACcESS.
11     #
12    
13     """
14     A simple framework to run benchmarks under OPENMP and to summarize the results in tables for instance in HTML
15    
16     @var __author__: name of author
17     @var __licence__: licence agreement
18     var __url__: url entry point on documentation
19     @var __version__: version
20     @var __date__: date of the version
21     """
22    
23     __author__="Lutz Gross, l.gross@uq.edu.au"
24     __licence__="contact: esys@access.uq.edu.au"
25     __url__="http://www.iservo.edu.au/esys/escript"
26     __version__="$Revision:$"
27     __date__="$Date:$"
28    
29     import os,socket,time,sys
30    
31     class BenchmarkSuite(object):
32     """
33     framework to run a bunch of L{Benchmark}s with the object to create a table of statistics.
34     @var MAX_LEVEL: maximum number of level in headers for output
35     """
36     MAX_LEVEL=5
37     def __init__(self,name=None):
38     """
39     sets up a suite of benchmarks
40    
41     @param name: name of the benchmark suite. If no name is given the class name is used.
42     @type name: C{str}
43     """
44     super(BenchmarkSuite,self).__init__()
45     self.__benchmarks=[]
46     self.__scale=1
47     if name==None:
48     self.__name=self.__class__.__name__
49     else:
50     self.__name=name
51    
52     def __str__(self):
53     """
54     returns the name of the benchmark suite
55    
56     @return: name
57     @rtype: C{str}
58     """
59     return self.__name
60     def addBenchmark(self,benchmark):
61     """
62     adds a L{Benchmark} to the suite
63    
64     @param benchmark: adds a new L{Benchmark} to the suite
65     @type benchmark: L{Benchmark}
66     """
67     self.__benchmarks.append(benchmark)
68     def __len__(self):
69     """
70     returns the number of benchmarks in the suite
71    
72     @return: number of benchmarks
73     @rtype: C{int}
74     """
75     return len(self.__benchmarks)
76     def __getitem__(self,i):
77     """
78     returns the i-th benchmark in the suite through self[i]
79    
80     @param i: index of the requested benchmark
81     @type i: C{int}
82     @return: i-th benchmark
83     @rtype: L{Benchmark}
84    
85     """
86     return self.__benchmarks[i]
87     def run(self,scale=1):
88     """
89     runs all benchmarks
90    
91     @param scale: defines the number of (OpenMP) threads to be used. If scale is a scalar all benchmarks
92     are run with scale number of threads. If scale is a C{list}, the p-th problem in each of the benchmarks
93 gross 386 in the suite is run with scale[p] threads.
94 gross 364 @type scale: C{int} or C{list} of C{int}s.
95     """
96     self.__scale=scale
97 gross 379 for i in range(len(self)): self[i].run(scale=scale)
98     def getHTML(self,filter,level=1):
99 gross 364 """
100     returns the results of the last benchmark run in HTML format.
101    
102 gross 379 @param filter: filter to be applied to the results
103     @type filter: L{BenchmarkFilter}
104 gross 364 @param level: level used in header <H?> tags
105     @type level: C{int}
106     @return: HTML document
107     @rtype: C{str}
108     """
109     out=""
110     if level==1: out+="<HTML><HEAD><TITLE>Benchmark: %s</TITLE></HEAD><BODY>\n"%str(self)
111     out+="<H%s>%s</H%s>\n"%(level,str(self),level)
112     if level==1:
113     m=""
114     if isinstance(self.__scale,int):
115     if self.__scale>1:
116     m=" (%s threads)"%self.__scale
117     out+="<p>platform: %s%s</p>\n"%(socket.gethostname(),m)
118     for i in range(len(self)):
119     out+="<p>\n"
120 gross 379 out+=self[i].getHTML(filter=filter,level=min(level+1,self.MAX_LEVEL))
121 gross 364 out+="<p>\n"
122     if level==1:
123     out+="<hr><p align=\"center\">by %s at %s</p>\n"%(os.getlogin(),time.strftime('%X %x %Z'))
124     out+="</BODY></HTML>\n"
125     return out
126    
127    
128     class Benchmark(object):
129     """
130     runs a bunch of similar L{BenchmarkProblem}s with a bunch of L{Options}
131     """
132     def __init__(self,name=None,description=None):
133     """
134     sets up a benchmark
135    
136     @param name: name of the benchmark. If no name is given the class name is used.
137     @type name: C{str}
138     @param description: description of the benchmark.
139     @type description: C{str} or C{None}
140     """
141     super(Benchmark,self).__init__()
142     self.__options=[]
143     self.__problems=[]
144     self.__results=[]
145     self.__scale=1
146     if name==None:
147     self.__name=self.__class__.__name__
148     else:
149     self.__name=name
150     self.__description=description
151    
152     def __str__(self):
153     """
154     returns the name of the benchmark suite
155    
156     @return: name
157     @rtype: C{str}
158     """
159     return self.__name
160    
161     def addProblem(self,problem):
162     """
163     adds a problem to the benchmark
164    
165     @param problem: adds a new problem to the bechmark
166     @type problem: L{BenchmarkProblem}
167     """
168     self.__problems.append(problem)
169    
170     def addOptions(self,Options):
171     """
172     adds a options to the benchmark
173    
174     @param options: adds a new option to the bechmark
175     @type problem: L{Options}
176     """
177     self.__options.append(Options)
178    
179     def run(self,scale=1):
180     """
181     runs all problems with all options.
182    
183    
184     @param scale: defines the number of (OpenMP) threads to be used. If scale is a scalar all benchmarks
185     are run with scale number of threads. If scale is a C{list}, the p-th problem in each of the benchmarks
186 gross 386 in the suite is run with scale[p] threads.
187 gross 364 @type scale: C{int} or C{list} of C{int}s.
188     """
189     if isinstance(scale,list):
190 gross 386 c_max=min(len(scale),len(self.__problems))
191     else:
192     c_max=len(self.__problems)
193 gross 379 self.__filter=filter
194 gross 364 self.__scale=scale
195     self.__results=[]
196 gross 386 for c in range(c_max):
197     r=self.__problems[c]
198 gross 364 if isinstance(scale,list):
199     s=scale[c]
200     else:
201     s=scale
202     row=[]
203     for p in self.__options:
204 gross 385 os.environ['OMP_NUM_TREADS']=str(s)
205 gross 364 row.append(r.run(p))
206     self.__results.append(row)
207 gross 379 def getHTML(self,filter,level=1):
208 gross 364 """
209     returns the results of the last benchmark run in HTML format.
210    
211 gross 379 @param filter: filter to be applied to the results
212     @type filter: L{BenchmarkFilter}
213 gross 364 @param level: level used in header <H?> tags
214     @type level: C{int}
215     @return: HTML document
216     @rtype: C{str}
217     """
218     out=""
219     if level==1: out+="<HTML><HEAD><TITLE>Benchmark: %s</TITLE></HEAD><BODY>\n"%str(self)
220     out+="<H%s>%s</H%s>\n"%(level,str(self),level)
221     if level==1:
222     m=""
223     if isinstance(self.__scale,int):
224     if self.__scale>1:
225     m=" (%s threads)"%self.__scale
226     out+="<p>platform: %s%s</p>\n"%(socket.gethostname(),m)
227     if self.__description: out+="<p>%s</p>\n"%str(self.__description)
228     if len(self.__problems)>0:
229     out+="<TABLE ALIGN=\"center\" BORDER=3 CELLPADDING=5 CELLSPACING=1>\n"
230     h1_seg=""
231 gross 379 rn=filter.getResultNames()
232 gross 364 if len(rn)==0:
233     h1_seg+="<TD></TD>"
234     else:
235     for n in rn: h1_seg+="<TD ALIGN=\"center\">%s</TD>"%n
236     h0="<TR><TH ALIGN=\"center\" ROWSPAN=2>Case</TH>"
237     h1="<TR>"
238     if isinstance(self.__scale,list): h0+="<TH ALIGN=\"center\" ROWSPAN=2>Threads</TH>"
239     for o in self.__options:
240     if len(rn)==0:
241     h0+="<TH ALIGN=\"center\">%s</TH>"%str(o)
242     elif len(rn)==1:
243     h0+="<TH ALIGN=\"center\">%s</TH>"%str(o)
244     empty_h1=False
245     else:
246     h0+="<TH ALIGN=\"center\" COLSPAN=%s>%s</TH>"%(len(rn),str(o))
247     h1+=h1_seg
248     out+=h0+"</TR>\n"+h1+"</TR>\n"
249     c=0
250     for r in range(len(self.__results)):
251     out+="<TR><TH ALIGN=\"right\">%s</TH>"%str(self.__problems[r])
252     if isinstance(self.__scale,list): out+="<TD ALIGN=\"right\">%s</TD>"%self.__scale[c]
253 gross 385 for col in self.__results[r]:
254     for e in filter(col): out+="<TD ALIGN=\"right\">%s</TD>"%e
255 gross 364 out+="</TR>\n"
256 gross 385 c+=1
257 gross 364 out+="</TABLE>"
258     if level==1:
259     out+="<hr><p align=\"center\">by %s at %s</p>\n"%(os.getlogin(),time.strftime('%X %x %Z'))
260     out+="</BODY></HTML>\n"
261     return out
262    
263     class BenchmarkProblem(object):
264     """
265     something that can be run and returns a list of characteristics such as timing, Mflops, error, etc.
266     """
267     def __init__(self,name=None):
268     """
269     sets up a benchmark problem
270    
271     @param name: name of the problem. If no name is given the class name is used.
272     @type name: C{str}
273     """
274     super(BenchmarkProblem,self).__init__()
275     if name==None:
276     self.__name=self.__class__.__name__
277     else:
278     self.__name=name
279    
280    
281     def __str__(self):
282     """
283     returns the name of the benchmark suite
284    
285     @return: name
286     @rtype: C{str}
287     """
288     return self.__name
289    
290     def run(self,options=None):
291     """
292     runs the problem and returns a list of run characteristics
293    
294    
295     @param options: the options that are used for the run. Note that the number of OpenMP threads is controlled
296     by the L{Benchmark} the problem is run in.
297     @type options: L{Options}
298 gross 379 @return: run characteristics
299     @rtype: any type that can be read by the L{BenchmarkFilter} applied to it.
300 gross 364 @remark: this function has to overwritten by a particular problem
301     """
302     raise NotImplementedError
303     return []
304    
305 gross 379 class BenchmarkFilter(object):
306     """
307     object to filter the characteristcs returned by Bechmark runs.
308    
309     """
310     def __init__(self):
311     """
312     sets up a filter
313     """
314     pass
315    
316    
317     def getResultNames(self):
318     """
319     return the names of the results produced when run() is called.
320    
321     @return: names the list of the names to be used when the results of the run() call are printed
322     @rtype: C{list} of C{str}
323     @remark: this function has to overwritten by a particular problem
324     """
325     raise NotImplementedError
326     return []
327    
328     def __call__(self,result):
329     """
330     filters out values results returned as characteristcs of a problem run
331    
332     @param result: values to be filtered
333     @type result: any type that is produced by the L{BenchmarkProblem} it is applied to
334     @return: a list of strings selected from result
335     @rtype: C{list} of C{str}
336     @remark: this function has to overwritten by a particular problem
337     """
338     raise NotImplementedError
339     return []
340    
341    
342 gross 364 class Options(object):
343     """
344     defines a set of options to be used to run a L{BenchmarkProblem}
345     """
346     def __init__(self,name=None):
347     """
348     sets up the options
349    
350     @param name: name of the option. If no name is given the class name is used.
351     @type name: C{str}
352     """
353     super(Options,self).__init__()
354     if name==None:
355 gross 385 self.__name=self.__class__.__name__
356 gross 364 else:
357     self.__name=name
358     def __str__(self):
359     """
360     returns the name of the benchmark suite
361    
362     @return: name
363     @rtype: C{str}
364     """
365     return self.__name
366    
367     if __name__=="__main__":
368    
369     class OptionsTest1(Options):
370     pass
371     class OptionsTest2(Options):
372     pass
373    
374 gross 379 class BenchmarkProblemTest1(BenchmarkProblem):
375 gross 364 def __init__(self):
376     super(BenchmarkProblemTest1,self).__init__(name="TEST1")
377     def run(self,options=None):
378     import time
379     return time.time(),"A"
380    
381 gross 379 class BenchmarkProblemTest2(BenchmarkProblem):
382 gross 364 def __init__(self):
383     super(BenchmarkProblemTest2,self).__init__(name="TEST2")
384     def run(self,options=None):
385     import time
386     return -time.time(),"B"
387    
388 gross 379 class SimpleFilter(BenchmarkFilter):
389     def getResultNames(self):
390     return ["r0","r1"]
391     def __call__(self,result):
392     return [str(result[0]),str(result[1])]
393    
394 gross 364 bm=Benchmark("Example")
395     bm.addProblem(BenchmarkProblemTest1())
396     bm.addProblem(BenchmarkProblemTest2())
397     bm.addOptions(OptionsTest1())
398     bm.addOptions(OptionsTest2())
399    
400     bms=BenchmarkSuite("A Test")
401     bms.addBenchmark(bm)
402    
403     bms.run()
404 gross 379 print bms.getHTML(filter=SimpleFilter())
405 gross 364
406     bms.run(scale=4)
407 gross 379 print bms.getHTML(filter=SimpleFilter())
408 gross 364
409     bms.run(scale=[1,2])
410 gross 379 print bms.getHTML(filter=SimpleFilter())

  ViewVC Help
Powered by ViewVC 1.1.26