> Here's a question: suppose we've got a database result with 10K rows (I'd > say that is fairly common), and we're processing each row with a regex > (something that can't be done in SQL). What's a ballpark for gc overhead > before and after your fix? (I'm still not set up to compile CVS, so I > can't do it myself.) For 10K objects, the GC overhead is negligeable. Jeremy did 100K tuples and still found that 60% of the time was in malloc. You only get in trouble when you approach a million tuples. Remember, it's quadratic. That means it gets 100x worse with every multiplying factor 10 -- but also that it gets 100x better with every division by 10 (until you run into other effects, of course). --Guido van Rossum (home page: http://www.python.org/~guido/)
RetroSearch is an open source project built by @garambo | Open a GitHub Issue
Search and Browse the WWW like it's 1997 | Search results from DuckDuckGo
HTML:
3.2
| Encoding:
UTF-8
| Version:
0.7.4