timeline.go 5.4 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233
  1. // Package parser provides Timeline parsing based on timeline.jison
  2. package parser
  3. import (
  4. "fmt"
  5. "strings"
  6. "mermaid-go/pkg/ast"
  7. "mermaid-go/pkg/lexer"
  8. )
  9. // TimelineParser implements Timeline parsing following timeline.jison
  10. type TimelineParser struct {
  11. tokens []lexer.Token
  12. current int
  13. diagram *ast.TimelineDiagram
  14. }
  15. // NewTimelineParser creates a new Timeline parser
  16. func NewTimelineParser() *TimelineParser {
  17. return &TimelineParser{
  18. diagram: &ast.TimelineDiagram{
  19. Sections: make([]*ast.TimelineSection, 0),
  20. Config: make(map[string]any),
  21. },
  22. }
  23. }
  24. // Parse parses Timeline syntax
  25. func (p *TimelineParser) Parse(input string) (*ast.TimelineDiagram, error) {
  26. // Tokenize
  27. l := lexer.NewLexer(input)
  28. tokens, err := l.Tokenize()
  29. if err != nil {
  30. return nil, fmt.Errorf("lexical analysis failed: %w", err)
  31. }
  32. // Filter tokens
  33. p.tokens = lexer.FilterTokens(tokens)
  34. p.current = 0
  35. p.diagram = &ast.TimelineDiagram{
  36. Sections: make([]*ast.TimelineSection, 0),
  37. Config: make(map[string]any),
  38. }
  39. // Parse document
  40. err = p.parseDocument()
  41. if err != nil {
  42. return nil, fmt.Errorf("syntax analysis failed: %w", err)
  43. }
  44. return p.diagram, nil
  45. }
  46. // parseDocument parses the Timeline document
  47. func (p *TimelineParser) parseDocument() error {
  48. // Expect timeline
  49. if !p.check(lexer.TokenID) || p.peek().Value != "timeline" {
  50. return p.error("expected 'timeline'")
  51. }
  52. p.advance()
  53. // Parse statements
  54. for !p.isAtEnd() {
  55. if err := p.parseStatement(); err != nil {
  56. return err
  57. }
  58. }
  59. return nil
  60. }
  61. // parseStatement parses individual Timeline statements
  62. func (p *TimelineParser) parseStatement() error {
  63. if p.isAtEnd() {
  64. return nil
  65. }
  66. switch {
  67. case p.check(lexer.TokenNewline):
  68. p.advance() // Skip newlines
  69. return nil
  70. case p.checkKeyword("title"):
  71. return p.parseTitle()
  72. case p.checkKeyword("section"):
  73. return p.parseSection()
  74. case p.check(lexer.TokenColon):
  75. return p.parseEvent()
  76. case p.check(lexer.TokenID):
  77. // Period definition
  78. return p.parsePeriod()
  79. default:
  80. token := p.peek()
  81. return p.error(fmt.Sprintf("unexpected token: %s", token.Value))
  82. }
  83. }
  84. // parseTitle parses title statements
  85. func (p *TimelineParser) parseTitle() error {
  86. p.advance() // consume 'title'
  87. var titleParts []string
  88. for !p.check(lexer.TokenNewline) && !p.isAtEnd() {
  89. titleParts = append(titleParts, p.advance().Value)
  90. }
  91. if len(titleParts) > 0 {
  92. title := strings.TrimSpace(strings.Join(titleParts, " "))
  93. p.diagram.Title = &title
  94. }
  95. return nil
  96. }
  97. // parseSection parses section statements
  98. func (p *TimelineParser) parseSection() error {
  99. p.advance() // consume 'section'
  100. var sectionParts []string
  101. for !p.check(lexer.TokenNewline) && !p.isAtEnd() {
  102. sectionParts = append(sectionParts, p.advance().Value)
  103. }
  104. if len(sectionParts) > 0 {
  105. sectionName := strings.TrimSpace(strings.Join(sectionParts, " "))
  106. section := &ast.TimelineSection{
  107. Name: sectionName,
  108. Events: make([]*ast.TimelineEvent, 0),
  109. }
  110. p.diagram.Sections = append(p.diagram.Sections, section)
  111. }
  112. return nil
  113. }
  114. // parseEvent parses event statements (starting with :)
  115. func (p *TimelineParser) parseEvent() error {
  116. p.advance() // consume ':'
  117. var eventParts []string
  118. for !p.check(lexer.TokenNewline) && !p.isAtEnd() {
  119. eventParts = append(eventParts, p.advance().Value)
  120. }
  121. if len(eventParts) > 0 {
  122. eventText := strings.TrimSpace(strings.Join(eventParts, " "))
  123. event := &ast.TimelineEvent{
  124. Name: eventText,
  125. }
  126. // Add to current section or create default section
  127. if len(p.diagram.Sections) == 0 {
  128. section := &ast.TimelineSection{
  129. Name: "Timeline",
  130. Events: make([]*ast.TimelineEvent, 0),
  131. }
  132. p.diagram.Sections = append(p.diagram.Sections, section)
  133. }
  134. currentSection := p.diagram.Sections[len(p.diagram.Sections)-1]
  135. currentSection.Events = append(currentSection.Events, event)
  136. }
  137. return nil
  138. }
  139. // parsePeriod parses period statements
  140. func (p *TimelineParser) parsePeriod() error {
  141. var periodParts []string
  142. for !p.check(lexer.TokenColon) && !p.check(lexer.TokenNewline) && !p.isAtEnd() {
  143. periodParts = append(periodParts, p.advance().Value)
  144. }
  145. if len(periodParts) > 0 {
  146. periodName := strings.TrimSpace(strings.Join(periodParts, " "))
  147. // Create a section for this period
  148. section := &ast.TimelineSection{
  149. Name: periodName,
  150. Events: make([]*ast.TimelineEvent, 0),
  151. }
  152. p.diagram.Sections = append(p.diagram.Sections, section)
  153. }
  154. return nil
  155. }
  156. // Helper methods
  157. func (p *TimelineParser) check(tokenType lexer.TokenType) bool {
  158. if p.isAtEnd() {
  159. return false
  160. }
  161. return p.peek().Type == tokenType
  162. }
  163. func (p *TimelineParser) checkKeyword(keyword string) bool {
  164. if p.isAtEnd() {
  165. return false
  166. }
  167. token := p.peek()
  168. return token.Type == lexer.TokenID && strings.ToLower(token.Value) == strings.ToLower(keyword)
  169. }
  170. func (p *TimelineParser) advance() lexer.Token {
  171. if !p.isAtEnd() {
  172. p.current++
  173. }
  174. return p.previous()
  175. }
  176. func (p *TimelineParser) isAtEnd() bool {
  177. return p.current >= len(p.tokens) || p.peek().Type == lexer.TokenEOF
  178. }
  179. func (p *TimelineParser) peek() lexer.Token {
  180. if p.current >= len(p.tokens) {
  181. return lexer.Token{Type: lexer.TokenEOF}
  182. }
  183. return p.tokens[p.current]
  184. }
  185. func (p *TimelineParser) previous() lexer.Token {
  186. if p.current <= 0 {
  187. return lexer.Token{Type: lexer.TokenEOF}
  188. }
  189. return p.tokens[p.current-1]
  190. }
  191. func (p *TimelineParser) error(message string) error {
  192. token := p.peek()
  193. return fmt.Errorf("parse error at line %d, column %d: %s (got %s)",
  194. token.Line, token.Column, message, token.Type.String())
  195. }