"Stephen J. Turnbull" <stephen@xemacs.org> writes: > I don't see any need for a deviation of the implementation from the > spec. You probably haven't looked at the code of the Python parser, either. > Just slurp in the whole file in the specified encoding. Then cast > the Unicode characters in ordinary literal strings down to bytesize It's not that simple. Or, perhaps, it is - but still somebody needs to write this. I won't find the time for a stage 2 implementation anytime soon, but I still would like to see the feature in Python 2.3. Even without looking at the parser code, you find two alternative implementations. Trust me that you will find more alternatives when you start writing the parser, and more problems. There is a number of aspects that need to be preserved. Performance is one of them, usage of the tokenizer for pgen is another. Regards, Martin
RetroSearch is an open source project built by @garambo | Open a GitHub Issue
Search and Browse the WWW like it's 1997 | Search results from DuckDuckGo
HTML:
3.2
| Encoding:
UTF-8
| Version:
0.7.4