Mercurial > traipse
view plugins/cherrypy/lib/filter/cachefilter.py @ 39:ed322725b928 ornery-orc tip
Traipse 'OpenRPG' {110114-00}
Traipse is a distribution of OpenRPG that is designed to be easy to setup and go. Traipse also makes it easy for developers to work on code without fear of sacrifice. 'Ornery-Orc' continues the trend of 'Grumpy' and adds fixes to the code. 'Ornery-Orc's main goal is to offer more advanced features and enhance the productivity of the user.
Update Summary (Closed)
New Features:
New to Map, can re-order Grid, Miniatures, and Whiteboard layer draw order
New to Server GUI, can now clear log
New Earthdawn Dieroller
New IronClaw roller, sheet, and image
New ShapeShifter PC Sheet
Updates:
Update to Warhammer PC Sheet. Rollers set as macros. Should work with little maintanence.
Update to Browser Server window. Display rooms with ' " & cleaner
Update to Server. Handles ' " & cleaner
Update to Dieroller. Cleaner, more effecient expression system
Update to Hidden Die plugin, allows for non standard dice rolls
Update to location.py, allows for more portable references when starting Traipse
Update to the Features node
Fixes:
Fix to InterParse that was causing an Infernal Loop with Namespace Internal
Fix to XML data, removed old Minidom and switched to Element Tree
Fix to Server that was causing eternal attempt to find a Server ID, in Register Rooms thread
Fix to Server, removing wxPython dependencies where not needed
Fix to metaservers.xml file not being created
Fix to Single and Double quotes in Whiteboard text
Fix to Background images not showing when using the Image Server
Fix to Duplicate chat names appearing
Fix to Server GUI's logging output
Fix to FNB.COLORFUL_TABS bug
Fix to Gametree for XSLT Sheets
Fix to Gametree for locating gametree files
Fix to Send to Chat from Gametree
Fix to Gametree, renaming and remapping operates correctly
Fix to aliaslib, prevents error caused when SafeHTML is sent None
author | sirebral |
---|---|
date | Fri, 14 Jan 2011 05:24:52 -0600 |
parents | 4385a7d0efd1 |
children |
line wrap: on
line source
""" Copyright (c) 2004, CherryPy Team (team@cherrypy.org) All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: * Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer. * Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. * Neither the name of the CherryPy Team nor the names of its contributors may be used to endorse or promote products derived from this software without specific prior written permission. THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. """ import threading import Queue import time import cStringIO from basefilter import BaseInputFilter, RequestHandled from cherrypy import cpg def defaultCacheKey(): return cpg.request.browserUrl class Tee: """ Wraps a stream object; chains the content that is written and keep a copy in a StringIO for caching purposes. """ def __init__(self, wfile, maxobjsize): self.wfile = wfile self.cache = cStringIO.StringIO() self.maxobjsize = maxobjsize self.caching = True self.size = 0 def write(self, s): self.wfile.write(s) if self.caching: self.size += len(s) if self.size < self.maxobjsize: self.cache.write(s) else: # exceeded the limit, aborts caching self.stopCaching() def flush(self): self.wfile.flush() def close(self): self.wfile.close() if self.caching: self.stopCaching() def stopCaching(self): self.caching = False self.cache.close() class MemoryCache: def __init__(self, key, delay, maxobjsize, maxsize, maxobjects): self.key = key self.delay = delay self.maxobjsize = maxobjsize self.maxsize = maxsize self.maxobjects = maxobjects self.cursize = 0 self.cache = {} self.expirationQueue = Queue.Queue() self.expirationThread = threading.Thread(target=self.expireCache, name='expireCache') self.expirationThread.setDaemon(True) self.expirationThread.start() self.totPuts = 0 self.totGets = 0 self.totHits = 0 self.totExpires = 0 self.totNonModified = 0 def expireCache(self): while True: expirationTime, objSize, objKey = self.expirationQueue.get(block=True, timeout=None) while (time.time() < expirationTime): time.sleep(0.1) try: del self.cache[objKey] self.totExpires += 1 self.cursize -= objSize except KeyError: # the key may have been deleted elsewhere pass def get(self): """ If the content is in the cache, returns a tuple containing the expiration time, the lastModified response header and the object (rendered as a string); returns None if the key is not found. """ self.totGets += 1 cacheItem = self.cache.get(self.key(), None) if cacheItem: self.totHits += 1 return cacheItem else: return None def put(self, lastModified, obj): objSize = len(obj) totalSize = self.cursize + objSize # checks if there's space for the object if ((objSize < self.maxobjsize) and (totalSize < self.maxsize) and (len(self.cache) < self.maxobjects)): # add to the expirationQueue & cache try: expirationTime = time.time() + self.delay objKey = self.key() self.expirationQueue.put((expirationTime, objSize, objKey)) self.totPuts += 1 self.cursize += objSize except Queue.Full: # can't add because the queue is full return self.cache[objKey] = (expirationTime, lastModified, obj) class CacheInputFilter(BaseInputFilter): """ Works on the input chain. If the page is already stored in the cache serves the contents. If the page is not in the cache, it wraps the cpg.response.wfile object; in this way, everything that is written is recorded, independent if it was sent directly or not. """ def __init__( self, CacheClass=MemoryCache, key=defaultCacheKey, delay=600, # 10 minutes maxobjsize=100000, # 100 KB maxsize=10000000, # 10 MB maxobjects=1000 # 1000 objects ): cpg._cache = CacheClass(key, delay, maxobjsize, maxsize, maxobjects) def afterRequestBody(self): """ Checks if the page is already in the cache """ cacheData = cpg._cache.get() if cacheData: expirationTime, lastModified, obj = cacheData # found a hit! check the if-modified-since request header modifiedSince = cpg.request.headerMap.get('If-Modified-Since', None) #print "Cache hit: If-Modified-Since=%s, lastModified=%s" % (modifiedSince, lastModified) if modifiedSince == lastModified: cpg._cache.totNonModified += 1 # the code below was borrowed from the sendResponse function # it should be refactored & put into a function to allow reuse cpg.response.wfile.write('%s %s\r\n' % (cpg.configOption.protocolVersion, 304)) # the code below doesn't work because the data isn't available at this point... #cpg.response.wfile.write('%s: %s\r\n' % ('Date', cpg.request.headerMap['Date'])) # should the cache record & replay cookies it too? cpg.response.wfile.write('\r\n') raise RequestHandled else: # serve it & get out from the request cpg.response.wfile.write(obj) raise RequestHandled else: # sets a wrapper to cache the contents cpg.response.wfile = Tee(cpg.response.wfile, cpg._cache.maxobjsize) cpg.threadData.cacheable = True class CacheOutputFilter(object): """ Works on the output chain. Stores the content of the page in the cache. """ def beforeResponse(self): """ Checks if the page is cacheable; if not so disables the cache. Uses a flag that may be reset by intermediate filters. Note that the output filter is usually the last filter in the chain, so this method is probably the last one called before the response is written. """ if isinstance(cpg.response.wfile, Tee): if cpg.threadData.cacheable: return # cancel caching wrapper = cpg.response.wfile wrapper.stopCaching() cpg.response.wfile = wrapper.wfile def afterResponse(self): """ Close & fix the cache entry after content was fully written """ if isinstance(cpg.response.wfile, Tee): wrapper = cpg.response.wfile if wrapper.caching: if cpg.response.headerMap.get('Pragma', None) != 'no-cache': lastModified = cpg.response.headerMap.get('Last-Modified', None) # saves the cache data cpg._cache.put(lastModified, wrapper.cache.getvalue()) # closes the wrapper wrapper.stopCaching() cpg.response.wfile = wrapper.wfile def percentual(n,d): """calculates the percentual, dealing with div by zeros""" if d == 0: return 0 else: return (float(n)/float(d))*100 def formatSize(n): """formats a number as a memory size, in bytes, kbytes, MB, GB)""" if n < 1024: return "%4d bytes" % n elif n < 1024*1024: return "%4d kbytes" % (n / 1024) elif n < 1024*1024*1024: return "%4d MB" % (n / (1024*1024)) else: return "%4d GB" % (n / (1024*1024*1024)) class CacheStats: def index(self): cpg.response.headerMap['Content-Type'] = 'text/plain' cpg.response.headerMap['Pragma'] = 'no-cache' cache = cpg._cache yield "Cache statistics\n" yield "Maximum object size: %s\n" % formatSize(cache.maxobjsize) yield "Maximum cache size: %s\n" % formatSize(cache.maxsize) yield "Maximum number of objects: %d\n" % cache.maxobjects yield "Current cache size: %s\n" % formatSize(cache.cursize) yield "Approximated expiration queue size: %d\n" % cache.expirationQueue.qsize() yield "Number of cache entries: %d\n" % len(cache.cache) yield "Total cache writes: %d\n" % cache.totPuts yield "Total cache read attempts: %d\n" % cache.totGets yield "Total hits: %d (%1.2f%%)\n" % (cache.totHits, percentual(cache.totHits, cache.totGets)) yield "Total misses: %d (%1.2f%%)\n" % (cache.totGets-cache.totHits, percentual(cache.totGets-cache.totHits, cache.totGets)) yield "Total expires: %d\n" % cache.totExpires yield "Total non-modified content: %d\n" % cache.totNonModified index.exposed = True