You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Implementing a lexer for PDF is essential for efficiently parsing and analyzing PDF documents. Unlike traditional programming languages, PDF documents have a unique structure and encoding, requiring a specialized lexer to interpret the document's syntax and content accurately.
By creating a dedicated lexer, we can ensure more precise parsing, improve performance, and enhance the maintainability of the PDF codebase, ultimately leading to a better user experience in handling and displaying PDF files.
Acceptance criteria
The lexer should handle all PDF types and tokens, according to the ISO 32000-1:2008 specification
The lexer should include the trivia, used later for Lossless syntax trees
The lexer should have error handling
Minimal performance improvements are required as part of this story
Rust does not attach the whitespace characters as trivia to tokens, but there was a discussion on that. Rust Analyzer however is implemented like Roslyn or Swift
Background and Motivation
Implementing a lexer for PDF is essential for efficiently parsing and analyzing PDF documents. Unlike traditional programming languages, PDF documents have a unique structure and encoding, requiring a specialized lexer to interpret the document's syntax and content accurately.
By creating a dedicated lexer, we can ensure more precise parsing, improve performance, and enhance the maintainability of the PDF codebase, ultimately leading to a better user experience in handling and displaying PDF files.
Acceptance criteria
ISO 32000-1:2008
specificationOpen questions
Resources
The text was updated successfully, but these errors were encountered: