wiki:DeveloperGuidelines/Optimisation

Version 82 (modified by Michael Howden, 12 years ago) ( diff )

--

Optimisation

Writing code which runs fast.

NB Performance is not everything, and you should not optimize any code without actually having a problem with it. In most cases, response times of around 600ms are totally acceptable (or a minor compared with the average download times on-site).

There are more things to consider besides performance, e.g. usability and maintainability of code. New contributors should be able to easily understand the code, and bugs should be easy to find.

Tips

Python

JavaScript

JavaScript string concatenation:

When to load the scripts:

Useful walk-through performance from Google Maps team:

"Slow post-load response is more harmful to user satisfaction than slow page load times, according to current HCI research."

Web2Py

  1. Optimize the models, throw away what we don't need. Every field counts.
    Especially problematic in view of performance are references (joins), as they execute implicit DB requests. The more complex references are, the slower the model loads.
  1. Function definitions in models do _NOT_ harm - the functions are not executed when the model is loaded, but just compiled - pre-compilation of the whole application gives a speed-up of just 10ms (compare to the total execution times!).
    In contrast to that, module-level commands (which are executed at time of loading of the model, e.g. CRUD string definitions) slow it down. Suggestion: put them into a "config" function for that model, and call only as needed.
  1. Everything that is static should be in the "static" folder - that applies especially to static JavaScript. If you put such in views, it gets processed by the view compiler and passed as dynamic content, which is totally unnecessary. Loading from static also gives the advantage that it gets cached by the webserver _and_ the client.
  1. Avoid _implicit_ redirects! (that is, without user interaction, e.g. as in open_module. There may be redirects that cannot be avoided.).
    A redirect simply doubles the response time (executes a new request and thus loads it all again).
  1. Be careful with Ajax - this might work nicely in local environments, but in real-world deployments this has shown to be unreliable and slow.
  1. Consider having configurations which are read from DB frequently but written-to rarely, be set in configuration files which are written-out from the DB (like the CSS from themes)
  1. Mind your field representations! Sometimes they execute complex DB queries, which behave fine with just a few test records, but become a nightmare with 10000 records. To be sure - test loading a form/list view with 10000 records in the table.
  1. Be careful with how you user Storages & Rows:
For Variable A B (optimized) Time A Time B
x = Storage(a="abc") x.a xa 1.064 0.39
x = Storage() x.a = "b" xa = "b" 0.49 0.24
row = !s3db.pr_person[1] !rowfirst_name row.first_name 0.25 4.18
table = s3db.pr_person "first_name" in table hasattr(table, "first_name") 0.74 0.42
table = s3db.pr_person "first_name" in table table.first_name 0.74 0.27

XSLT

  1. Traversing the tree (//-selectors!) is very slow. Indexes (<xsl:key) can improve performance up to 10000%
  1. Explicit <xsl:for-each> can be significant slower than template matching. Consider a separate template!

Specific Examples

Python

NB These vary on cases, so use the Profiler (and argument -F profiler.log when running web2py.py) to see how they work in your cases...

for i in xrange(0, len(rows)):
     row = rows[i]

runs much faster than:

 for row in rows:

(0.05 vs. 0.001 seconds in one test case, 2x improvement in another & a slight negative improvement in a 3rd).


value = db(table.id == id).select(table.field, limitby=(0, 1)).first()

runs 1.5x faster than:

value = table[id].field

(0.012 vs. 0.007 seconds vs in a test case)

NB If only expecting one record then the limitby provides a big speedup!


dict.get("x")

is significantly slower than

dict["x"]

but...the construction:

y = dict.get("x", None)
   if y:

is as fast as

if x in dict:
   y = dict["x"]

Also, keys() is slowing things down.

if a in dict.keys()

is ~25% slower than:

if a in dict

Another thing is that the profiler showed that there is extensive use of isinstance. So I tried to find an alternative, which would be:

if type(x) == "yyy"

In fact, this is ~30% faster than isinstance, but it won't find subclasses. So, if you test for:

if isinstance(x, dict)

and want Storages to match, then you cannot replace isinstance.

A real killer is hasattr(). I ran 5 million loops of

if "a" in dict:

vs.

if hasattr(dict, "a")

which was 4.5s vs. 12s.

Hence - for dicts, avoid hasattr to test for containment.

Golden Rules for DB Queries

These "rules" might seem a matter of course, however, sometimes you need to take a second look at your code:

  • Insert a temporary print >> sys.stderr, self.query into web2py's select() function and take a look at what it says.

Use Joins

One complex query is usually more efficient than multiple simple queries (and gives the DB server a chance to optimize):

records = db(db.mytable.name == name).select()
for r in records:
    other_records = db(db.othertable.code == r.code).select()

better:

rows = db((db.mytable.name == name) & (db.othertable.code == db.mytable.code)).select()
for row in rows:
    mytable_record = row.mytable
    othertable_record = row.othertable

Limit your Query

Ask exactly for what you expect - if you expect only one result, then limit the search by limitby:

db(db.mytable.id == id).select().first()

should be:

db(db.mytable.id == id).select(limitby=(0,1)).first()

If you need only certain fields of a record, then don't ask for all:

my_value = db(db.mytable.id == id).select(limitby=(0,1)).first().value

should be:

my_value = db(db.mytable.id == id).select(db.mytable.value, limitby=(0,1)).first().value

Don't ask twice…

...for the same record. Look down your code whether you need the same record again later:

my_value = db(db.mytable.id == id).select(db.mytable.value, limitby=(0,1)).first().value
...
other_value = db(db.mytable.id == id).select(db.mytable.other_value, limitby=(0,1)).first().other_value

better:

row = db(db.mytable.id == id).select(db.mytable.value, db.mytable.other_value, limitby=(0,1)).first()
if row:
    my_value = row.value
    other_value = row.other_value

Don't loop over Queries

...if you can avoid it:

for id in ids:
   my_record = db(db.mytable.id == id).select().first()
   ...

(much) better:

records = db(db.mytable.id.belongs(ids)).select()
for record in records:
   ...

Sometimes it is not as easy to see as in the above example - it could be hidden:

for x in y:
  id = some_function(x)
  if id:
    record = db(db.mytable.id == id).select()

better:

ids = filter(lambda x: some_function(x), y)
if ids:
    records = db(db.mytable.id.belongs(ids)).select()
    for record in records:
        ...

Or more complex:

for x in y:
    if x.a == some_value:
        record = db(db.mytable.id == x.key).select()
        ...<branch 1>
    else:
        record = db(db.othertable.id == x.other_key).select()
        ...<branch 2>

could be:

ids1 = filter(lambda x: (x.a == some_value) and x.key or None, y)
ids2 = filter(lambda x: (x.a != some_value) and x.other_key or None, y)
if ids1:
    records = db(db.mytable.id.belongs(ids1)).select()
    for record in records:
        ...<branch 1>
if ids2:
    records = db(db.othertable.id.belongs(ids2)).select()
    for record in records:
        ...<branch 2>

Profiling

HTTP Packet Sizes & Uplinks:

Simple script to place in a controller:

def test():
    from datetime import datetime
    start = datetime.today()
    for i in xrange(0, 1000):
        <testcode: variant A>
    middle = datetime.datetime.today()
    for i in xrange(0, 1000):
        <testcode: variant B>
    end = datetime.today()
    a = middle - start
    b = end - middle
    output = TABLE(TR("A: %s" % a),
                   TR("B: %s" % b))
    return output.xml()

Scalability

Much of your code will behave well with just a few test records in the database.

But in real-world scenarios, the system might have to handle tens of thousands of records - which means that your functions could be called tens of thousands of times in a single request. Will it still give reasonable response-times?

Ah - so you think that your onvalidation function will be called only once after the user has submitted the form, and so it doesn't harm that it takes 50ms compared with 600ms for the rest of the request? Maybe you're wrong: in an XML import of 100.000 records that same onvalidation routine would be called 100.000 times in a single request, which would give 1 hour 23 minutes for your onvalidation function alone, compared to 600ms for the rest of the request!

Also be careful in representation functions, especially when used in validators (IS_ONE_OF!). If the table contains thousands of records, then the representation will be called many many times to build a single form field (autocompletes can help!).

Always think in scales of a few thousand records, especially for sequentially called routines like representation of field values, onvalidation and onaccept functions. Do you really need to do all this once per record, or could it be done just once in the request?


DeveloperGuidelines

Note: See TracWiki for help on using the wiki.