-
-
Notifications
You must be signed in to change notification settings - Fork 31.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Improve tokenize error handling #104169
Comments
pablogsal
added a commit
to pablogsal/cpython
that referenced
this issue
May 4, 2023
pablogsal
added a commit
that referenced
this issue
May 4, 2023
CC: @lysnikolaou |
lysnikolaou
added a commit
to lysnikolaou/cpython
that referenced
this issue
Oct 11, 2023
* The lexer, which include the actual lexeme producing logic, goes into the `lexer` directory. * The wrappers, one wrapper per input mode (file, string, utf-8, and readline), go into the `tokenizer` directory and include logic for creating a lexer instance and managing the buffer for different modes.
lysnikolaou
added a commit
that referenced
this issue
Oct 11, 2023
* The lexer, which include the actual lexeme producing logic, goes into the `lexer` directory. * The wrappers, one wrapper per input mode (file, string, utf-8, and readline), go into the `tokenizer` directory and include logic for creating a lexer instance and managing the buffer for different modes. --------- Co-authored-by: Pablo Galindo <[email protected]> Co-authored-by: blurb-it[bot] <43283697+blurb-it[bot]@users.noreply.github.com>
lysnikolaou
added a commit
to lysnikolaou/cpython
that referenced
this issue
Oct 11, 2023
lysnikolaou
added a commit
that referenced
this issue
Oct 12, 2023
* Fix test_peg_generator after tokenizer refactoring * Remove references to tokenizer.c in comments etc.
Ready to close? Thanks! |
Glyphack
pushed a commit
to Glyphack/cpython
that referenced
this issue
Sep 2, 2024
…10684) * The lexer, which include the actual lexeme producing logic, goes into the `lexer` directory. * The wrappers, one wrapper per input mode (file, string, utf-8, and readline), go into the `tokenizer` directory and include logic for creating a lexer instance and managing the buffer for different modes. --------- Co-authored-by: Pablo Galindo <[email protected]> Co-authored-by: blurb-it[bot] <43283697+blurb-it[bot]@users.noreply.github.com>
Glyphack
pushed a commit
to Glyphack/cpython
that referenced
this issue
Sep 2, 2024
…ython#110727) * Fix test_peg_generator after tokenizer refactoring * Remove references to tokenizer.c in comments etc.
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
There has been quite a lot of instances were poor error handling in the tokenizer leads to crashes or errors being overwritten. We should not rely on continuous patches to every single individual issue but we should improve the situation with more robust infrastructure.
Linked PRs
The text was updated successfully, but these errors were encountered: