盐城市葛武初级中学“音乐之声”社团
社团的组建宗旨
20##年9月,我们学校成立了“音乐之声”社团,社团的成立
以“培养学生对音乐的兴趣爱好”为宗旨。社团开展各项活动以活跃
校园文化、激发同学自我展现、提升学校艺术底蕴、丰富学生课余生
活,为爱好音乐的同学们提供一个音乐学生与交流的平台。使学生在
每月的社团活动中得到充实的锻炼与音乐技能的学习,培养其创新精
神,致力构建学校的音乐文化,同时在乐器、声乐和合唱学习中培养
同学们的团队合作意识和责任感。并把这种责任感贯彻学习和生活之
中。
20##年9月
社团的机构设置
顾问 组长 指导教师
(卞青山) (顾善锟) (顾善锟)
社团管理章程
第一章 总 则
第一条:社团的性质
音乐之声社团,是一个开展音乐艺术实践和展示个人才华的学生性质的社团,是葛武初级中学在校学生音乐艺术骨干的团队。
第二条:社团的活动宗旨
1、 促进学校音乐艺术教育的发展,反映当代学生意气风发的精神面貌弘扬中华民族传统文化。
2、 丰富校园文化生活,活跃人文气氛,在艺术实践活动中进行美育和爱国主义教育,陶冶情操。
3、 提高学生个人文化艺术修养和审美观,建立一支艺术骨干队伍,使学校艺术活动上层次、上水平。
第三条:社团的任务
1、 训练及提高社团成员的专业水平。
2、 排练丰富多彩的节目,活跃校园文化生活。
3、 积极参与学校重大文艺演出及其他社区艺术活动。
第二章 组织机构
1、社团采取学生指导,教师辅助的形式。设社长一名,副社长两名,负责社团的各方面工作;各其他部设分社长一名,负责本组日常训练和成员管理。
2、社团团长、副团长及各分社长为本社团的骨干成员。
3、每学期初通过选举,重新选出团长、副团长及各分社长。
第三章 权利和义务
1、参加训练和接受培训的权利;提出合理化意见和建议的权利;选举和被选举的权利;有自由选择多种团体的权利。
2、团员有遵守本团章程,按时参加本团训练、演出,遵守本社团决议和各项规章制度,维护本团利益的义务。
3、不迟到、不早退、不缺勤,有特殊情况必须提前请假,爱护社团的公共财物;为社团作宣传,吸纳新成员。
第四章 规章制度
1、凡被音乐社团录用的成员,社团将统一建立成员档案,记载训练、演出、比赛及奖惩情况。
2、准时参加培训,不迟到、不早退、不缺勤,否则,给予处分。
3、各分社训练和演出由各分社长负责考勤,指导老师签字确认。
4、除特殊情况外,不能按时参加训练的必须事先请假,否则以旷课论处。
5、鼓励积极参加专业活动的积极分子和刻苦训练、表现突出的分社和个人。
6、每次训练结束,由分社长安排社员对场地进行清理打扫,保证场地干净整洁。
第五章 活动形式和内容
社团的活动采取定期训练的方式,活动时间为每周三和周五下午课外活动时间,若有演出任务,增加训练,时间安排临时通知。
第六章 附则
本章程的修改由音乐社团骨干讨论,经校政教处审查后实行。本章程自公布之日起生效
盐城市葛武初级中学“音乐之声”社团
20##年9月
社团活动计划
1、社团活动课,做到“四定”。即定内容、定人员、定时间、定
地点,发展学生的个性特色,以此推动学生音乐素质的全面发展。
2、完成平常音乐课堂教学的同时,不断扩大学生的音乐视野,
丰富学生的精神生活,进一步培养、发展学生的音乐兴趣,使他们学
有所长,培养学生的集体主义精神。
3、常督促学生完成一些学习音乐的好习惯,并使之不断巩固、
加强。
4、对小组的课程做出规范,以提高学生学习兴趣,培养音乐能
力为主要内容,不让学生放任自由。
5、培养学生的特长,并在一定的时间让他们发挥自己的特长,锻炼和发展学生的能力
本学期主要活动安排
第二篇:音乐之声鉴赏
profile.doc last updated 6/23/94 [by Guido]
PROFILER DOCUMENTATION and (mini) USER'S MANUAL
Copyright 1994, by InfoSeek Corporation, all rights reserved. Written by James Roskind
Permission to use, copy, modify, and distribute this Python software and its associated documentation for any purpose (subject to the
restriction in the following sentence) without fee is hereby granted, provided that the above copyright notice appears in all copies, and that both that copyright notice and this permission notice appear in supporting documentation, and that the name of InfoSeek not be used in advertising or publicity pertaining to distribution of the software without specific, written prior permission. This permission is
explicitly restricted to the copying and modification of the software to remain in Python, compiled Python, or other languages (such as C) wherein the modified or derived code is exclusively imported into a Python module.
INFOSEEK CORPORATION DISCLAIMS ALL WARRANTIES WITH REGARD TO THIS SOFTWARE, INCLUDING ALL IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL INFOSEEK CORPORATION BE LIABLE FOR ANY SPECIAL, INDIRECT OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
The profiler was written after only programming in Python for 3 weeks. As a result, it is probably clumsy code, but I don't know for sure yet 'cause I'm a beginner :-). I did work hard to make the code run fast, so that profiling would be a reasonable thing to do. I tried not to repeat code fragments, but I'm sure I did some stuff in really awkward ways at times. Please send suggestions for improvements to:
jar@infoseek.com. I won't promise *any* support. ...but I'd appreciate the feedback.
SECTION HEADING LIST:
INTRODUCTION
HOW IS THIS profile DIFFERENT FROM THE OLD profile MODULE?
INSTANT USERS MANUAL
WHAT IS DETERMINISTIC PROFILING?
REFERENCE MANUAL
FUNCTION profile.run(string, filename_opt)
CLASS Stats(filename, ...)
METHOD strip_dirs()
METHOD add(filename, ...)
METHOD sort_stats(key, ...)
METHOD reverse_order()
METHOD print_stats(restriction, ...)
METHOD print_callers(restrictions, ...)
METHOD print_callees(restrictions, ...)
METHOD ignore()
LIMITATIONS
CALIBRATION
EXTENSIONS: Deriving Better Profilers
INTRODUCTION
A "profiler" is a program that describes the run time performance of a program, providing a variety of statistics. This documentation describes the profiler functionality provided in the modules
"profile" and "pstats." This profiler provides "deterministic profiling" of any Python programs. It also provides a series of
report generation tools to allow users to rapidly examine the results of a profile operation.
HOW IS THIS profile DIFFERENT FROM THE OLD profile MODULE?
The big changes from standard profiling module are that you get more information, and you pay less CPU time. It's not a trade-off, it's a trade-up.
To be specific:
bugs removed: local stack frame is no longer molested, execution time is now charged to correct functions, ....
accuracy increased: profiler execution time is no longer charged to user's code, calibration for platform is supported, file reads are not done *by* profiler *during* profiling (and charged to user's code!), ...
speed increased: Overhead CPU cost was reduced by more than a factor of two (perhaps a factor of five), lightweight profiler module is all that must be loaded, and the report generating module
(pstats) is not needed during profiling.
recursive functions support: cumulative times in recursive functions are correctly calculated; recursive entries are counted; ...
large growth in report generating UI: distinct profiles runs can be added together forming a comprehensive report; functions that import statistics take arbitrary lists of files; sorting criteria is now based on keywords (instead of 4 integer options); reports shows what functions were profiled as well as what profile file was referenced; output format has been improved, ...
INSTANT USERS MANUAL
This section is provided for users that "don't want to read the
manual." It provides a very brief overview, and allows a user to
rapidly perform profiling on an existing application.
To profile an application with a main entry point of "foo()", you would add the following to your module:
import profile profile.run("foo()")
The above action would cause "foo()" to be run, and a series of
informative lines (the profile) to be printed. The above approach is most useful when working with the interpreter. If you would like to save the results of a profile into a file for later examination, you can supply a file name as the second argument to the run() function:
When you wish to review the profile, you should use the methods in the pstats module. Typically you would load the statistics data as
follows:
import pstats p = pstats.Stats('fooprof') import profile profile.run("foo()", 'fooprof')
The class "Stats" (the above code just created an instance of this
class) has a variety of methods for manipulating and printing the data that was just read into "p". When you ran profile.run() above, what was printed was the result of three method calls:
p.strip_dirs().sort_stats(-1).print_stats()
The first method removed the extraneous path from all the module names. The second method sorted all the entries according to the standard module/line/name string that is printed (this is to comply with the semantics of the old profiler). The third method printed out all the statistics. You might try the following sort calls:
p.sort_stats('name') p.print_stats()
The first call will actually sort the list by function name, and the second call will print out the statistics. The following are some interesting calls to experiment with:
p.sort_stats('cumulative').print_stats(10)
This sorts the profile by cumulative time in a function, and then only prints the ten most significant lines. If you want to understand what algorithms are taking time, the above line is what you would use.
If you were looking to see what functions were looping a lot, and taking a lot of time, you would do:
p.sort_stats('time').print_stats(10)
to sort according to time spent within each function, and then print the statistics for the top ten functions.
You might also try:
p.sort_stats('file').print_stats('__init__')
This will sort all the statistics by file name, and then print out statistics for only the class init methods ('cause they are spelled with "__init__" in them). As one final example, you could try:
p.sort_stats('time', 'cum').print_stats(.5, 'init')
This line sorts stats with a primary key of time, and a secondary key of cumulative time, and then prints out some of the statistics. To be specific, the list is first culled down to 50% (re: .5) of its
original size, then only lines containing "init" are maintained, and that sub-sub-list is printed.
If you wondered what functions called the above functions, you could now (p is still sorted according to the last criteria) do:
and you would get a list of callers for each of the listed functions.
If you want more functionality, you're going to have to read the manual (or guess) what the following functions do:
p.print_callees() p.add('fooprof') p.print_callers(.5, 'init')
WHAT IS DETERMINISTIC PROFILING?
"Deterministic profiling" is meant to reflect the fact that all "function call", "function return", and "exception" events are monitored, and precise timings are made for the intervals between these events (during which time the user's code is executing). In contrast, "statistical profiling" (which is not done by this module) randomly samples the effective instruction pointer, and deduces where time is being spent. The latter technique traditionally involves less overhead (as the code does not need to be instrumented), but provides only relative indications of where time is being spent.
In Python, since there is an interpreter active during execution, the presence of instrumented code is not required to do deterministic profiling. Python automatically provides a hook (optional callback) for each event. In addition, the interpreted nature of Python tends to add so much overhead to execution, that deterministic profiling tends to only add small processing overhead, in typical applications. The result is that deterministic profiling is not that expensive, but yet provides extensive run time statistics about the execution of a Python program.
Call count statistics can be used to identify bugs in code (surprising counts), and to identify possible inline-expansion points (high call
counts). Internal time statistics can be used to identify hot loops that should be carefully optimized. Cumulative time statistics should be used to identify high level errors in the selection of algorithms. Note that the unusual handling of cumulative times in this profiler allows statistics for recursive implementations of algorithms to be directly compared to iterative implementations.
REFERENCE MANUAL
The primary entry point for the profiler is the global function
profile.run(). It is typically used to create any profile
information. The reports are formatted and printed using methods for the class pstats.Stats. The following is a description of all of these standard entry points and functions. For a more in-depth view of some of the code, consider reading the later section on "Profiler Extensions," which includes discussion of how to derive "better"
profilers from the classes presented, or reading the source code for these modules.
FUNCTION profile.run(string, filename_opt)
This function takes a single argument that has can be passed to the "exec" statement, and an optional file name. In all cases this
routine attempts to "exec" its first argument, and gather profiling statistics from the execution. If no file name is present, then this function automatically prints a simple profiling report, sorted by the standard name string (file/line/function-name) that is presented in each line. The following is a typical output from such a call:
cut here----
main()
2706 function calls (2004 primitive calls) in 4.504 CPU seconds
Ordered by: standard name
ncalls tottime percall cumtime percall filename:lineno(function) 2 0.006 0.003 0.953 0.477 pobject.py:75(save_objects) 43/3 0.533 0.012 0.749 0.250 pobject.py:99(evaluate) ...
cut here----
The first line indicates that this profile was generated by the call: profile.run('main()'), and hence the exec'ed string is 'main()'. The second line indicates that 2706 calls were monitored. Of those calls, 2004 were "primitive." We define "primitive" to mean that the call was not induced via recursion. The next line: "Ordered by: standard name", indicates that the text string in the far right column was used to sort the output. The column headings include:
"ncalls" for the number of calls, "tottime" for the total time spent in the given function (and excluding time made in calls to sub-functions), "percall" is the quotient of "tottime" divided by "ncalls" "cumtime" is the total time spent in this and all subfunctions (i.e., from invocation till exit). This figure is accurate *even* for recursive functions. "percall" is the quotient of "cumtime" divided by primitive calls "filename:lineno(function)" provides the respective data of each function
When there are two numbers in the first column (e.g.: 43/3), then the latter is the number of primitive calls, and the former is the actual number of calls. Note that when the function does not recurse, these two values are the same, and only the single figure is printed.
CLASS Stats(filename, ...)
This class constructor creates an instance of a statistics object from a filename (or set of filenames). Stats objects are manipulated by methods, in order to print useful reports.
The file selected by the above constructor must have been created by the corresponding version of profile. To be specific, there is *NO* file compatibility guaranteed with future versions of this profiler, and there is no compatibility with files produced by other profilers (e.g., the standard system profiler).
If several files are provided, all the statistics for identical functions will be coalesced, so that an overall view of several
processes can be considered in a single report. If additional files need to be combined with data in an existing Stats object, the add() method can be used.
METHOD strip_dirs()
This method for the Stats class removes all leading path information from file names. It is very useful in reducing the size of the
printout to fit within (close to) 80 columns. This method modifies the object, and the striped information is lost. After performing a strip operation, the object is considered to have its entries in a "random" order, as it was just after object initialization and loading. If strip_dir() causes two function names to be
indistinguishable (i.e., they are on the same line of the same
filename, and have the same function name), then the statistics for these two entries are accumulated into a single entry.
METHOD add(filename, ...)
This methods of the Stats class accumulates additional profiling
information into the current profiling object. Its arguments should refer to filenames created my the corresponding version of
profile.run(). Statistics for identically named (re: file, line, name) functions are automatically accumulated into single function statistics.
METHOD sort_stats(key, ...)
This method modifies the Stats object by sorting it according to the supplied criteria. The argument is typically a string identifying the basis of a sort (example: "time" or "name").
When more than one key is provided, then additional keys are used as secondary criteria when the there is equality in all keys selected before them. For example, sort_stats('name', 'file') will sort all the entries according to their function name, and resolve all ties (identical function names) by sorting by file name.
Abbreviations can be used for any key names, as long as the
abbreviation is unambiguous. The following are the keys currently defined:
Valid Arg Meaning "calls" call count
"cumulative" cumulative time "file" file name "module" file name "pcalls" primitive call count "line" line number "name" function name "nfl" name/file/line "stdname" standard name "time" internal time
Note that all sorts on statistics are in descending order (placing most time consuming items first), where as name, file, and line number searches are in ascending order (i.e., alphabetical). The subtle
distinction between "nfl" and "stdname" is that the standard name is a sort of the name as printed, which means that the embedded line
numbers get compared in an odd way. For example, lines 3, 20, and 40 would (if the file names were the same) appear in the string order "20" "3" and "40". In contrast, "nfl" does a numeric compare of the line numbers. In fact, sort_stats("nfl") is the same as
sort_stats("name", "file", "line").
For compatibility with the standard profiler, the numeric argument -1, 0, 1, and 2 are permitted. They are interpreted as "stdname",
"calls", "time", and "cumulative" respectively. If this old style format (numeric) is used, only one sort key (the numeric key) will be used, and additionally arguments will be silently ignored.
METHOD reverse_order()
This method for the Stats class reverses the ordering of the basic list within the object. This method is provided primarily for
compatibility with the standard profiler. Its utility is questionable now that ascending vs descending order is properly selected based on the sort key of choice.
METHOD print_stats(restriction, ...)
This method for the Stats class prints out a report as described in the profile.run() definition.
The order of the printing is based on the last sort_stats() operation done on the object (subject to caveats in add() and strip_dirs()).
The arguments provided (if any) can be used to limit the list down to the significant entries. Initially, the list is taken to be the complete set of profiled functions. Each restriction is either an integer (to select a count of lines), or a decimal fraction between 0.0 and 1.0 inclusive (to select a percentage of lines), or a regular expression (to pattern match the standard name that is printed). If several restrictions are provided, then they are applied sequentially. For example:
would first limit the printing to first 10% of list, and then only print functions that were part of filename ".*foo:". In contrast, the command:
print_stats("foo:", .1)
would limit the list to all functions having file names ".*foo:", and then proceed to only print the first 10% of them.
METHOD print_callers(restrictions, ...)
This method for the Stats class prints a list of all functions that called each function in the profiled database. The ordering is
identical to that provided by print_stats(), and the definition of the restricting argument is also identical. For convenience, a number is shown in parentheses after each caller to show how many times this specific call was made. A second non-parenthesized number is the cumulative time spent in the function at the right.
METHOD print_callees(restrictions, ...)
This method for the Stats class prints a list of all function that were called by the indicated function. Aside from this reversal of direction of calls (re: called vs was called by), the arguments and ordering are identical to the print_callers() method.
METHOD ignore()
This method of the Stats class is used to dispose of the value print_stats(.1, "foo:")
returned by earlier methods. All standard methods in this class
return the instance that is being processed, so that the commands can be strung together. For example:
pstats.Stats('foofile').strip_dirs().sort_stats('cum').print_stats().ignore()
would perform all the indicated functions, but it would not return
the final reference to the Stats instance.
LIMITATIONS
There are two fundamental limitations on this profiler. The first is that it relies on the Python interpreter to dispatch "call", "return", and "exception" events. Compiled C code does not get interpreted,
and hence is "invisible" to the profiler. All time spent in C code
(including builtin functions) will be charged to the Python function that was invoked the C code. IF the C code calls out to some native Python code, then those calls will be profiled properly.
The second limitation has to do with accuracy of timing information. There is a fundamental problem with deterministic profilers involving accuracy. The most obvious restriction is that the underlying "clock" is only ticking at a rate (typically) of about .001 seconds. Hence no measurements will be more accurate that that underlying clock. If
enough measurements are taken, then the "error" will tend to average out. Unfortunately, removing this first error induces a second source of error...
The second problem is that it "takes a while" from when an event is
dispatched until the profiler's call to get the time actually *gets* the state of the clock. Similarly, there is a certain lag when
exiting the profiler event handler from the time that the clock's
value was obtained (and then squirreled away), until the user's code is once again executing. As a result, functions that are called many times, or call many functions, will typically accumulate this error. The error that accumulates in this fashion is typically less than the accuracy of the clock (i.e., less than one clock tick), but it *can* accumulate and become very significant. This profiler provides a
means of calibrating itself for a give platform so that this error can be probabilistically (i.e., on the average) removed. After the
profiler is calibrated, it will be more accurate (in a least square
sense), but it will sometimes produce negative numbers (when call
counts are exceptionally low, and the gods of probability work against you :-). ) Do *NOT* be alarmed by negative numbers in the profile. They should *only* appear if you have calibrated your profiler, and the results are actually better than without calibration.
CALIBRATION
The profiler class has a hard coded constant that is added to each
event handling time to compensate for the overhead of calling the time function, and socking away the results. The following procedure can be used to obtain this constant for a given platform (see discussion in LIMITATIONS above).
import profile pr = profile.Profile() pr.calibrate(100) pr.calibrate(100) pr.calibrate(100)
The argument to calibrate() is the number of times to try to do the sample calls to get the CPU times. If your computer is *very* fast, you might have to do:
pr.calibrate(1000)
or even:
pr.calibrate(10000)
The object of this exercise is to get a fairly consistent result.
When you have a consistent answer, you are ready to use that number in the source code. For a Sun Sparcstation 1000 running Solaris 2.3, the magical number is about .00053. If you have a choice, you are better off with a smaller constant, and your results will "less often" show up as negative in profile statistics.
The following shows how the trace_dispatch() method in the Profile class should be modified to install the calibration constant on a Sun Sparcstation 1000:
def trace_dispatch(self, frame, event, arg): t = self.timer()
t = t[0] + t[1] - self.t - .00053 # Calibration constant if self.dispatch[event](frame,t): t = self.timer() self.t = t[0] + t[1] else: r = self.timer() self.t = r[0] + r[1] - t # put back unrecorded delta return
Note that if there is no calibration constant, then the line
containing the callibration constant should simply say:
t = t[0] + t[1] - self.t # no calibration constant
You can also achieve the same results using a derived class (and the profiler will actually run equally fast!!), but the above method is the simplest to use. I could have made the profiler "self
calibrating", but it would have made the initialization of the profiler class slower, and would have required some *very* fancy coding, or else the use of a variable where the constant .00053 was placed in the code shown. This is a ****VERY**** critical performance section, and there is no reason to use a variable lookup at this point, when a constant can be used.
EXTENSIONS: Deriving Better Profilers
The Profile class of profile was written so that derived classes
could be developed to extend the profiler. Rather than describing all the details of such an effort, I'll just present the following two examples of derived classes that can be used to do profiling. If the reader is an avid Python programmer, then it should be possible to use these as a model and create similar (and perchance better) profile classes.
If all you want to do is change how the timer is called, or which
timer function is used, then the basic class has an option for that in the constructor for the class. Consider passing the name of a function to call into the constructor:
pr = profile.Profile(your_time_func)
The resulting profiler will call your time function instead of
os.times(). The function should return either a single number, or a
list of numbers (like what os.times() returns). If the function
returns a single time number, or the list of returned numbers has
length 2, then you will get an especially fast version of the dispatch routine.
Be warned that you *should* calibrate the profiler class for the
timer function that you choose. For most machines, a timer that
returns a lone integer value will provide the best results in terms of low overhead during profiling. (os.times is *pretty* bad, 'cause it
returns a tuple of floating point values, so all arithmetic is
floating point in the profiler!). If you want to be substitute a
better timer in the cleanest fashion, you should derive a class, and
simply put in the replacement dispatch method that better handles your timer call, along with the appropriate calibration constant :-).
cut here------------------------------------------------------------------ #**************************************************************************** # OldProfile class documentation
#**************************************************************************** #
# The following derived profiler simulates the old style profile, providing # errant results on recursive functions. The reason for the usefulness of this # profiler is that it runs faster (i.e., less overhead) than the old
# profiler. It still creates all the caller stats, and is quite
# useful when there is *no* recursion in the user's code. It is also # a lot more accurate than the old profiler, as it does not charge all # its overhead time to the user's code.
#**************************************************************************** class OldProfile(Profile):
def trace_dispatch_exception(self, frame, t):
def trace_dispatch_call(self, frame, t): fn = `frame.f_code` self.cur = (t, 0, 0, fn, frame, self.cur) if self.timings.has_key(fn): tt, ct, callers = self.timings[fn] self.timings[fn] = tt, ct, callers rt, rtt, rct, rfn, rframe, rcur = self.cur if rcur and not rframe is frame: return self.trace_dispatch_return(rframe, t) return 0
else: self.timings[fn] = 0, 0, {} return 1 def trace_dispatch_return(self, frame, t): rt, rtt, rct, rfn, frame, rcur = self.cur rtt = rtt + t sft = rtt + rct pt, ptt, pct, pfn, pframe, pcur = rcur self.cur = pt, ptt+rt, pct+sft, pfn, pframe, pcur tt, ct, callers = self.timings[rfn] if callers.has_key(pfn): callers[pfn] = callers[pfn] + 1 else: callers[pfn] = 1 self.timings[rfn] = tt+rtt, ct + sft, callers return 1 def snapshot_stats(self): self.stats = {} for func in self.timings.keys(): tt, ct, callers = self.timings[func] nor_func = self.func_normalize(func) nor_callers = {} nc = 0 for func_caller in callers.keys(): nor_callers[self.func_normalize(func_caller)]=\ callers[func_caller] nc = nc + callers[func_caller] self.stats[nor_func] = nc, nc, tt, ct, nor_callers
#**************************************************************************** # HotProfile class documentation
#**************************************************************************** #
# This profiler is the fastest derived profile example. It does not
# calculate caller-callee relationships, and does not calculate cumulative # time under a function. It only calculates time spent in a function, so
# it runs very quickly (re: very low overhead). In truth, the basic # profiler is so fast, that is probably not worth the savings to give # up the data, but this class still provides a nice example.
#**************************************************************************** class HotProfile(Profile):
def snapshot_stats(self): self.stats = {} for func in self.timings.keys(): nc, tt = self.timings[func] nor_func = self.func_normalize(func) self.stats[nor_func] = nc, nc, tt, 0, {} def trace_dispatch_call(self, frame, t): self.cur = (t, 0, frame, self.cur) return 1 def trace_dispatch_exception(self, frame, t): rt, rtt, rfn, rframe, rcur = self.cur if rcur and not rframe is frame: return self.trace_dispatch_return(rframe, t) return 0 def trace_dispatch_return(self, frame, t): rt, rtt, frame, rcur = self.cur rfn = `frame.f_code` pt, ptt, pframe, pcur = rcur self.cur = pt, ptt+rt, pframe, pcur if self.timings.has_key(rfn): nc, tt = self.timings[rfn] self.timings[rfn] = nc + 1, rt + rtt + tt else: self.timings[rfn] = 1, rt + rtt return 1
cut here------------------------------------------------------------------