1use crate::{Grammar, InjectionConfig, Language, LanguageRegistry};
2use lazy_static::lazy_static;
3use parking_lot::Mutex;
4use std::{
5 borrow::Cow,
6 cell::RefCell,
7 cmp::{Ordering, Reverse},
8 collections::BinaryHeap,
9 ops::{Deref, DerefMut, Range},
10 sync::Arc,
11};
12use sum_tree::{Bias, SeekTarget, SumTree};
13use text::{rope, Anchor, BufferSnapshot, OffsetRangeExt, Point, Rope, ToOffset, ToPoint};
14use tree_sitter::{
15 Node, Parser, Query, QueryCapture, QueryCaptures, QueryCursor, QueryMatches, Tree,
16};
17
18thread_local! {
19 static PARSER: RefCell<Parser> = RefCell::new(Parser::new());
20}
21
22lazy_static! {
23 static ref QUERY_CURSORS: Mutex<Vec<QueryCursor>> = Default::default();
24}
25
26#[derive(Default)]
27pub struct SyntaxMap {
28 parsed_version: clock::Global,
29 interpolated_version: clock::Global,
30 snapshot: SyntaxSnapshot,
31 language_registry: Option<Arc<LanguageRegistry>>,
32}
33
34#[derive(Clone, Default)]
35pub struct SyntaxSnapshot {
36 layers: SumTree<SyntaxLayer>,
37}
38
39#[derive(Default)]
40pub struct SyntaxMapCaptures<'a> {
41 layers: Vec<SyntaxMapCapturesLayer<'a>>,
42 active_layer_count: usize,
43 grammars: Vec<&'a Grammar>,
44}
45
46#[derive(Default)]
47pub struct SyntaxMapMatches<'a> {
48 layers: Vec<SyntaxMapMatchesLayer<'a>>,
49 active_layer_count: usize,
50 grammars: Vec<&'a Grammar>,
51}
52
53#[derive(Debug)]
54pub struct SyntaxMapCapture<'a> {
55 pub depth: usize,
56 pub node: Node<'a>,
57 pub index: u32,
58 pub grammar_index: usize,
59}
60
61#[derive(Debug)]
62pub struct SyntaxMapMatch<'a> {
63 pub depth: usize,
64 pub pattern_index: usize,
65 pub captures: &'a [QueryCapture<'a>],
66 pub grammar_index: usize,
67}
68
69struct SyntaxMapCapturesLayer<'a> {
70 depth: usize,
71 captures: QueryCaptures<'a, 'a, TextProvider<'a>>,
72 next_capture: Option<QueryCapture<'a>>,
73 grammar_index: usize,
74 _query_cursor: QueryCursorHandle,
75}
76
77struct SyntaxMapMatchesLayer<'a> {
78 depth: usize,
79 next_pattern_index: usize,
80 next_captures: Vec<QueryCapture<'a>>,
81 has_next: bool,
82 matches: QueryMatches<'a, 'a, TextProvider<'a>>,
83 grammar_index: usize,
84 _query_cursor: QueryCursorHandle,
85}
86
87#[derive(Clone)]
88struct SyntaxLayer {
89 depth: usize,
90 range: Range<Anchor>,
91 tree: tree_sitter::Tree,
92 language: Arc<Language>,
93}
94
95#[derive(Debug)]
96pub struct SyntaxLayerInfo<'a> {
97 pub depth: usize,
98 pub node: Node<'a>,
99 pub language: &'a Arc<Language>,
100}
101
102#[derive(Debug, Clone)]
103struct SyntaxLayerSummary {
104 min_depth: usize,
105 max_depth: usize,
106 range: Range<Anchor>,
107 last_layer_range: Range<Anchor>,
108}
109
110#[derive(Clone, Debug)]
111struct DepthAndRange(usize, Range<Anchor>);
112
113#[derive(Clone, Debug)]
114struct DepthAndMaxPosition(usize, Anchor);
115
116#[derive(Clone, Debug)]
117struct DepthAndRangeOrMaxPosition(DepthAndRange, DepthAndMaxPosition);
118
119struct ReparseStep {
120 depth: usize,
121 language: Arc<Language>,
122 ranges: Vec<tree_sitter::Range>,
123 range: Range<Anchor>,
124}
125
126#[derive(Debug, PartialEq, Eq)]
127struct ChangedRegion {
128 depth: usize,
129 range: Range<Anchor>,
130}
131
132#[derive(Default)]
133struct ChangeRegionSet(Vec<ChangedRegion>);
134
135struct TextProvider<'a>(&'a Rope);
136
137struct ByteChunks<'a>(rope::Chunks<'a>);
138
139struct QueryCursorHandle(Option<QueryCursor>);
140
141impl SyntaxMap {
142 pub fn new() -> Self {
143 Self::default()
144 }
145
146 pub fn set_language_registry(&mut self, registry: Arc<LanguageRegistry>) {
147 self.language_registry = Some(registry);
148 }
149
150 pub fn snapshot(&self) -> SyntaxSnapshot {
151 self.snapshot.clone()
152 }
153
154 pub fn language_registry(&self) -> Option<Arc<LanguageRegistry>> {
155 self.language_registry.clone()
156 }
157
158 pub fn parsed_version(&self) -> clock::Global {
159 self.parsed_version.clone()
160 }
161
162 pub fn interpolate(&mut self, text: &BufferSnapshot) {
163 self.snapshot.interpolate(&self.interpolated_version, text);
164 self.interpolated_version = text.version.clone();
165 }
166
167 #[cfg(test)]
168 pub fn reparse(&mut self, language: Arc<Language>, text: &BufferSnapshot) {
169 self.snapshot.reparse(
170 &self.parsed_version,
171 text,
172 self.language_registry.clone(),
173 language,
174 );
175 self.parsed_version = text.version.clone();
176 self.interpolated_version = text.version.clone();
177 }
178
179 pub fn did_parse(&mut self, snapshot: SyntaxSnapshot, version: clock::Global) {
180 self.interpolated_version = version.clone();
181 self.parsed_version = version;
182 self.snapshot = snapshot;
183 }
184
185 pub fn clear(&mut self) {
186 self.snapshot = SyntaxSnapshot::default();
187 }
188}
189
190impl SyntaxSnapshot {
191 pub fn is_empty(&self) -> bool {
192 self.layers.is_empty()
193 }
194
195 pub fn interpolate(&mut self, from_version: &clock::Global, text: &BufferSnapshot) {
196 let edits = text
197 .anchored_edits_since::<(usize, Point)>(&from_version)
198 .collect::<Vec<_>>();
199 if edits.is_empty() {
200 return;
201 }
202
203 let mut layers = SumTree::new();
204 let mut first_edit_ix_for_depth = 0;
205 let mut prev_depth = 0;
206 let mut cursor = self.layers.cursor::<SyntaxLayerSummary>();
207 cursor.next(text);
208
209 'outer: loop {
210 let depth = cursor.end(text).max_depth;
211 if depth > prev_depth {
212 first_edit_ix_for_depth = 0;
213 prev_depth = depth;
214 }
215
216 // Preserve any layers at this depth that precede the first edit.
217 if let Some((_, edit_range)) = edits.get(first_edit_ix_for_depth) {
218 let target = DepthAndMaxPosition(depth, edit_range.start);
219 if target.cmp(&cursor.start(), text).is_gt() {
220 let slice = cursor.slice(&target, Bias::Left, text);
221 layers.push_tree(slice, text);
222 }
223 }
224 // If this layer follows all of the edits, then preserve it and any
225 // subsequent layers at this same depth.
226 else if cursor.item().is_some() {
227 let slice = cursor.slice(
228 &DepthAndRange(depth + 1, Anchor::MIN..Anchor::MAX),
229 Bias::Left,
230 text,
231 );
232 layers.push_tree(slice, text);
233 continue;
234 };
235
236 let layer = if let Some(layer) = cursor.item() {
237 layer
238 } else {
239 break;
240 };
241 let (start_byte, start_point) = layer.range.start.summary::<(usize, Point)>(text);
242
243 // Ignore edits that end before the start of this layer, and don't consider them
244 // for any subsequent layers at this same depth.
245 loop {
246 if let Some((_, edit_range)) = edits.get(first_edit_ix_for_depth) {
247 if edit_range.end.cmp(&layer.range.start, text).is_le() {
248 first_edit_ix_for_depth += 1;
249 } else {
250 break;
251 }
252 } else {
253 continue 'outer;
254 }
255 }
256
257 let mut layer = layer.clone();
258 for (edit, edit_range) in &edits[first_edit_ix_for_depth..] {
259 // Ignore any edits that follow this layer.
260 if edit_range.start.cmp(&layer.range.end, text).is_ge() {
261 break;
262 }
263
264 // Apply any edits that intersect this layer to the layer's syntax tree.
265 let tree_edit = if edit_range.start.cmp(&layer.range.start, text).is_ge() {
266 tree_sitter::InputEdit {
267 start_byte: edit.new.start.0 - start_byte,
268 old_end_byte: edit.new.start.0 - start_byte
269 + (edit.old.end.0 - edit.old.start.0),
270 new_end_byte: edit.new.end.0 - start_byte,
271 start_position: (edit.new.start.1 - start_point).to_ts_point(),
272 old_end_position: (edit.new.start.1 - start_point
273 + (edit.old.end.1 - edit.old.start.1))
274 .to_ts_point(),
275 new_end_position: (edit.new.end.1 - start_point).to_ts_point(),
276 }
277 } else {
278 let node = layer.tree.root_node();
279 tree_sitter::InputEdit {
280 start_byte: 0,
281 old_end_byte: node.end_byte(),
282 new_end_byte: 0,
283 start_position: Default::default(),
284 old_end_position: node.end_position(),
285 new_end_position: Default::default(),
286 }
287 };
288
289 layer.tree.edit(&tree_edit);
290 }
291
292 debug_assert!(
293 layer.tree.root_node().end_byte() <= text.len(),
294 "tree's size {}, is larger than text size {}",
295 layer.tree.root_node().end_byte(),
296 text.len(),
297 );
298
299 layers.push(layer, text);
300 cursor.next(text);
301 }
302
303 layers.push_tree(cursor.suffix(&text), &text);
304 drop(cursor);
305 self.layers = layers;
306 }
307
308 pub fn reparse(
309 &mut self,
310 from_version: &clock::Global,
311 text: &BufferSnapshot,
312 registry: Option<Arc<LanguageRegistry>>,
313 language: Arc<Language>,
314 ) {
315 let edits = text.edits_since::<usize>(from_version).collect::<Vec<_>>();
316 let max_depth = self.layers.summary().max_depth;
317 let mut cursor = self.layers.cursor::<SyntaxLayerSummary>();
318 cursor.next(&text);
319 let mut layers = SumTree::new();
320
321 let mut changed_regions = ChangeRegionSet::default();
322 let mut queue = BinaryHeap::new();
323 queue.push(ReparseStep {
324 depth: 0,
325 language: language.clone(),
326 ranges: Vec::new(),
327 range: Anchor::MIN..Anchor::MAX,
328 });
329
330 loop {
331 let step = queue.pop();
332 let (depth, range) = if let Some(step) = &step {
333 (step.depth, step.range.clone())
334 } else {
335 (max_depth + 1, Anchor::MAX..Anchor::MAX)
336 };
337
338 let target = DepthAndRange(depth, range.clone());
339 let mut done = cursor.item().is_none();
340 while !done && target.cmp(&cursor.end(text), &text).is_gt() {
341 done = true;
342
343 let bounded_target =
344 DepthAndRangeOrMaxPosition(target.clone(), changed_regions.start_position());
345 if bounded_target.cmp(&cursor.start(), &text).is_gt() {
346 let slice = cursor.slice(&bounded_target, Bias::Left, text);
347 if !slice.is_empty() {
348 layers.push_tree(slice, &text);
349 if changed_regions.prune(cursor.end(text), text) {
350 done = false;
351 }
352 }
353 }
354
355 while target.cmp(&cursor.end(text), text).is_gt() {
356 let layer = if let Some(layer) = cursor.item() {
357 layer
358 } else {
359 break;
360 };
361
362 if changed_regions.intersects(&layer, text) {
363 changed_regions.insert(
364 ChangedRegion {
365 depth: layer.depth + 1,
366 range: layer.range.clone(),
367 },
368 text,
369 );
370 } else {
371 layers.push(layer.clone(), text);
372 }
373
374 cursor.next(text);
375 if changed_regions.prune(cursor.end(text), text) {
376 done = false;
377 }
378 }
379 }
380
381 let (ranges, language) = if let Some(step) = step {
382 (step.ranges, step.language)
383 } else {
384 break;
385 };
386
387 let start_point;
388 let start_byte;
389 let end_byte;
390 if let Some((first, last)) = ranges.first().zip(ranges.last()) {
391 start_point = first.start_point;
392 start_byte = first.start_byte;
393 end_byte = last.end_byte;
394 } else {
395 start_point = Point::zero().to_ts_point();
396 start_byte = 0;
397 end_byte = text.len();
398 };
399
400 let mut old_layer = cursor.item();
401 if let Some(layer) = old_layer {
402 if layer.range.to_offset(text) == (start_byte..end_byte) {
403 cursor.next(&text);
404 } else {
405 old_layer = None;
406 }
407 }
408
409 let grammar = if let Some(grammar) = language.grammar.as_deref() {
410 grammar
411 } else {
412 continue;
413 };
414
415 let tree;
416 let changed_ranges;
417 if let Some(old_layer) = old_layer {
418 tree = parse_text(
419 grammar,
420 text.as_rope(),
421 Some(old_layer.tree.clone()),
422 ranges,
423 );
424 changed_ranges = join_ranges(
425 edits
426 .iter()
427 .map(|e| e.new.clone())
428 .filter(|range| range.start < end_byte && range.end > start_byte),
429 old_layer
430 .tree
431 .changed_ranges(&tree)
432 .map(|r| start_byte + r.start_byte..start_byte + r.end_byte),
433 );
434 } else {
435 tree = parse_text(grammar, text.as_rope(), None, ranges);
436 changed_ranges = vec![start_byte..end_byte];
437 }
438
439 layers.push(
440 SyntaxLayer {
441 depth,
442 range,
443 tree: tree.clone(),
444 language: language.clone(),
445 },
446 &text,
447 );
448
449 if let (Some((config, registry)), false) = (
450 grammar.injection_config.as_ref().zip(registry.as_ref()),
451 changed_ranges.is_empty(),
452 ) {
453 let depth = depth + 1;
454 for range in &changed_ranges {
455 changed_regions.insert(
456 ChangedRegion {
457 depth,
458 range: text.anchor_before(range.start)..text.anchor_after(range.end),
459 },
460 text,
461 );
462 }
463 get_injections(
464 config,
465 text,
466 tree.root_node_with_offset(start_byte, start_point),
467 registry,
468 depth,
469 &changed_ranges,
470 &mut queue,
471 );
472 }
473 }
474
475 drop(cursor);
476 self.layers = layers;
477 }
478
479 pub fn single_tree_captures<'a>(
480 range: Range<usize>,
481 text: &'a Rope,
482 tree: &'a Tree,
483 language: &'a Arc<Language>,
484 query: fn(&Grammar) -> Option<&Query>,
485 ) -> SyntaxMapCaptures<'a> {
486 SyntaxMapCaptures::new(
487 range.clone(),
488 text,
489 [SyntaxLayerInfo {
490 language,
491 depth: 0,
492 node: tree.root_node(),
493 }]
494 .into_iter(),
495 query,
496 )
497 }
498
499 pub fn captures<'a>(
500 &'a self,
501 range: Range<usize>,
502 buffer: &'a BufferSnapshot,
503 query: fn(&Grammar) -> Option<&Query>,
504 ) -> SyntaxMapCaptures {
505 SyntaxMapCaptures::new(
506 range.clone(),
507 buffer.as_rope(),
508 self.layers_for_range(range, buffer).into_iter(),
509 query,
510 )
511 }
512
513 pub fn matches<'a>(
514 &'a self,
515 range: Range<usize>,
516 buffer: &'a BufferSnapshot,
517 query: fn(&Grammar) -> Option<&Query>,
518 ) -> SyntaxMapMatches {
519 SyntaxMapMatches::new(
520 range.clone(),
521 buffer.as_rope(),
522 self.layers_for_range(range, buffer).into_iter(),
523 query,
524 )
525 }
526
527 #[cfg(test)]
528 pub fn layers(&self, buffer: &BufferSnapshot) -> Vec<SyntaxLayerInfo> {
529 self.layers_for_range(0..buffer.len(), buffer)
530 }
531
532 pub fn layers_for_range<'a, T: ToOffset>(
533 &self,
534 range: Range<T>,
535 buffer: &BufferSnapshot,
536 ) -> Vec<SyntaxLayerInfo> {
537 let start = buffer.anchor_before(range.start.to_offset(buffer));
538 let end = buffer.anchor_after(range.end.to_offset(buffer));
539
540 let mut cursor = self.layers.filter::<_, ()>(|summary| {
541 if summary.max_depth > summary.min_depth {
542 true
543 } else {
544 let is_before_start = summary.range.end.cmp(&start, buffer).is_lt();
545 let is_after_end = summary.range.start.cmp(&end, buffer).is_gt();
546 !is_before_start && !is_after_end
547 }
548 });
549
550 let mut result = Vec::new();
551 cursor.next(buffer);
552 while let Some(layer) = cursor.item() {
553 result.push(SyntaxLayerInfo {
554 language: &layer.language,
555 depth: layer.depth,
556 node: layer.tree.root_node_with_offset(
557 layer.range.start.to_offset(buffer),
558 layer.range.start.to_point(buffer).to_ts_point(),
559 ),
560 });
561 cursor.next(buffer)
562 }
563
564 result
565 }
566}
567
568impl<'a> SyntaxMapCaptures<'a> {
569 fn new(
570 range: Range<usize>,
571 text: &'a Rope,
572 layers: impl Iterator<Item = SyntaxLayerInfo<'a>>,
573 query: fn(&Grammar) -> Option<&Query>,
574 ) -> Self {
575 let mut result = Self {
576 layers: Vec::new(),
577 grammars: Vec::new(),
578 active_layer_count: 0,
579 };
580 for SyntaxLayerInfo {
581 language,
582 depth,
583 node,
584 } in layers
585 {
586 let grammar = match &language.grammar {
587 Some(grammer) => grammer,
588 None => continue,
589 };
590 let query = match query(&grammar) {
591 Some(query) => query,
592 None => continue,
593 };
594
595 let mut query_cursor = QueryCursorHandle::new();
596
597 // TODO - add a Tree-sitter API to remove the need for this.
598 let cursor = unsafe {
599 std::mem::transmute::<_, &'static mut QueryCursor>(query_cursor.deref_mut())
600 };
601
602 cursor.set_byte_range(range.clone());
603 let captures = cursor.captures(query, node, TextProvider(text));
604 let grammar_index = result
605 .grammars
606 .iter()
607 .position(|g| g.id == grammar.id())
608 .unwrap_or_else(|| {
609 result.grammars.push(grammar);
610 result.grammars.len() - 1
611 });
612 let mut layer = SyntaxMapCapturesLayer {
613 depth,
614 grammar_index,
615 next_capture: None,
616 captures,
617 _query_cursor: query_cursor,
618 };
619
620 layer.advance();
621 if layer.next_capture.is_some() {
622 let key = layer.sort_key();
623 let ix = match result.layers[..result.active_layer_count]
624 .binary_search_by_key(&key, |layer| layer.sort_key())
625 {
626 Ok(ix) | Err(ix) => ix,
627 };
628 result.layers.insert(ix, layer);
629 result.active_layer_count += 1;
630 } else {
631 result.layers.push(layer);
632 }
633 }
634
635 result
636 }
637
638 pub fn grammars(&self) -> &[&'a Grammar] {
639 &self.grammars
640 }
641
642 pub fn peek(&self) -> Option<SyntaxMapCapture<'a>> {
643 let layer = self.layers[..self.active_layer_count].first()?;
644 let capture = layer.next_capture?;
645 Some(SyntaxMapCapture {
646 depth: layer.depth,
647 grammar_index: layer.grammar_index,
648 index: capture.index,
649 node: capture.node,
650 })
651 }
652
653 pub fn advance(&mut self) -> bool {
654 let layer = if let Some(layer) = self.layers[..self.active_layer_count].first_mut() {
655 layer
656 } else {
657 return false;
658 };
659
660 layer.advance();
661 if layer.next_capture.is_some() {
662 let key = layer.sort_key();
663 let i = 1 + self.layers[1..self.active_layer_count]
664 .iter()
665 .position(|later_layer| key < later_layer.sort_key())
666 .unwrap_or(self.active_layer_count - 1);
667 self.layers[0..i].rotate_left(1);
668 } else {
669 self.layers[0..self.active_layer_count].rotate_left(1);
670 self.active_layer_count -= 1;
671 }
672
673 true
674 }
675
676 pub fn set_byte_range(&mut self, range: Range<usize>) {
677 for layer in &mut self.layers {
678 layer.captures.set_byte_range(range.clone());
679 if let Some(capture) = &layer.next_capture {
680 if capture.node.end_byte() > range.start {
681 continue;
682 }
683 }
684 layer.advance();
685 }
686 self.layers.sort_unstable_by_key(|layer| layer.sort_key());
687 self.active_layer_count = self
688 .layers
689 .iter()
690 .position(|layer| layer.next_capture.is_none())
691 .unwrap_or(self.layers.len());
692 }
693}
694
695impl<'a> SyntaxMapMatches<'a> {
696 fn new(
697 range: Range<usize>,
698 text: &'a Rope,
699 layers: impl Iterator<Item = SyntaxLayerInfo<'a>>,
700 query: fn(&Grammar) -> Option<&Query>,
701 ) -> Self {
702 let mut result = Self::default();
703 for SyntaxLayerInfo {
704 language,
705 depth,
706 node,
707 } in layers
708 {
709 let grammar = match &language.grammar {
710 Some(grammer) => grammer,
711 None => continue,
712 };
713 let query = match query(&grammar) {
714 Some(query) => query,
715 None => continue,
716 };
717
718 let mut query_cursor = QueryCursorHandle::new();
719
720 // TODO - add a Tree-sitter API to remove the need for this.
721 let cursor = unsafe {
722 std::mem::transmute::<_, &'static mut QueryCursor>(query_cursor.deref_mut())
723 };
724
725 cursor.set_byte_range(range.clone());
726 let matches = cursor.matches(query, node, TextProvider(text));
727 let grammar_index = result
728 .grammars
729 .iter()
730 .position(|g| g.id == grammar.id())
731 .unwrap_or_else(|| {
732 result.grammars.push(grammar);
733 result.grammars.len() - 1
734 });
735 let mut layer = SyntaxMapMatchesLayer {
736 depth,
737 grammar_index,
738 matches,
739 next_pattern_index: 0,
740 next_captures: Vec::new(),
741 has_next: false,
742 _query_cursor: query_cursor,
743 };
744
745 layer.advance();
746 if layer.has_next {
747 let key = layer.sort_key();
748 let ix = match result.layers[..result.active_layer_count]
749 .binary_search_by_key(&key, |layer| layer.sort_key())
750 {
751 Ok(ix) | Err(ix) => ix,
752 };
753 result.layers.insert(ix, layer);
754 result.active_layer_count += 1;
755 } else {
756 result.layers.push(layer);
757 }
758 }
759 result
760 }
761
762 pub fn grammars(&self) -> &[&'a Grammar] {
763 &self.grammars
764 }
765
766 pub fn peek(&self) -> Option<SyntaxMapMatch> {
767 let layer = self.layers.first()?;
768 if !layer.has_next {
769 return None;
770 }
771 Some(SyntaxMapMatch {
772 depth: layer.depth,
773 grammar_index: layer.grammar_index,
774 pattern_index: layer.next_pattern_index,
775 captures: &layer.next_captures,
776 })
777 }
778
779 pub fn advance(&mut self) -> bool {
780 let layer = if let Some(layer) = self.layers.first_mut() {
781 layer
782 } else {
783 return false;
784 };
785
786 layer.advance();
787 if layer.has_next {
788 let key = layer.sort_key();
789 let i = 1 + self.layers[1..self.active_layer_count]
790 .iter()
791 .position(|later_layer| key < later_layer.sort_key())
792 .unwrap_or(self.active_layer_count - 1);
793 self.layers[0..i].rotate_left(1);
794 } else {
795 self.layers[0..self.active_layer_count].rotate_left(1);
796 self.active_layer_count -= 1;
797 }
798
799 true
800 }
801}
802
803impl<'a> SyntaxMapCapturesLayer<'a> {
804 fn advance(&mut self) {
805 self.next_capture = self.captures.next().map(|(mat, ix)| mat.captures[ix]);
806 }
807
808 fn sort_key(&self) -> (usize, Reverse<usize>, usize) {
809 if let Some(capture) = &self.next_capture {
810 let range = capture.node.byte_range();
811 (range.start, Reverse(range.end), self.depth)
812 } else {
813 (usize::MAX, Reverse(0), usize::MAX)
814 }
815 }
816}
817
818impl<'a> SyntaxMapMatchesLayer<'a> {
819 fn advance(&mut self) {
820 if let Some(mat) = self.matches.next() {
821 self.next_captures.clear();
822 self.next_captures.extend_from_slice(&mat.captures);
823 self.next_pattern_index = mat.pattern_index;
824 self.has_next = true;
825 } else {
826 self.has_next = false;
827 }
828 }
829
830 fn sort_key(&self) -> (usize, Reverse<usize>, usize) {
831 if self.has_next {
832 let captures = &self.next_captures;
833 if let Some((first, last)) = captures.first().zip(captures.last()) {
834 return (
835 first.node.start_byte(),
836 Reverse(last.node.end_byte()),
837 self.depth,
838 );
839 }
840 }
841 (usize::MAX, Reverse(0), usize::MAX)
842 }
843}
844
845impl<'a> Iterator for SyntaxMapCaptures<'a> {
846 type Item = SyntaxMapCapture<'a>;
847
848 fn next(&mut self) -> Option<Self::Item> {
849 let result = self.peek();
850 self.advance();
851 result
852 }
853}
854
855fn join_ranges(
856 a: impl Iterator<Item = Range<usize>>,
857 b: impl Iterator<Item = Range<usize>>,
858) -> Vec<Range<usize>> {
859 let mut result = Vec::<Range<usize>>::new();
860 let mut a = a.peekable();
861 let mut b = b.peekable();
862 loop {
863 let range = match (a.peek(), b.peek()) {
864 (Some(range_a), Some(range_b)) => {
865 if range_a.start < range_b.start {
866 a.next().unwrap()
867 } else {
868 b.next().unwrap()
869 }
870 }
871 (None, Some(_)) => b.next().unwrap(),
872 (Some(_), None) => a.next().unwrap(),
873 (None, None) => break,
874 };
875
876 if let Some(last) = result.last_mut() {
877 if range.start <= last.end {
878 last.end = last.end.max(range.end);
879 continue;
880 }
881 }
882 result.push(range);
883 }
884 result
885}
886
887fn parse_text(
888 grammar: &Grammar,
889 text: &Rope,
890 old_tree: Option<Tree>,
891 mut ranges: Vec<tree_sitter::Range>,
892) -> Tree {
893 let (start_byte, start_point) = ranges
894 .first()
895 .map(|range| (range.start_byte, Point::from_ts_point(range.start_point)))
896 .unwrap_or_default();
897
898 for range in &mut ranges {
899 range.start_byte -= start_byte;
900 range.end_byte -= start_byte;
901 range.start_point = (Point::from_ts_point(range.start_point) - start_point).to_ts_point();
902 range.end_point = (Point::from_ts_point(range.end_point) - start_point).to_ts_point();
903 }
904
905 PARSER.with(|parser| {
906 let mut parser = parser.borrow_mut();
907 let mut chunks = text.chunks_in_range(start_byte..text.len());
908 parser
909 .set_included_ranges(&ranges)
910 .expect("overlapping ranges");
911 parser
912 .set_language(grammar.ts_language)
913 .expect("incompatible grammar");
914 parser
915 .parse_with(
916 &mut move |offset, _| {
917 chunks.seek(start_byte + offset);
918 chunks.next().unwrap_or("").as_bytes()
919 },
920 old_tree.as_ref(),
921 )
922 .expect("invalid language")
923 })
924}
925
926fn get_injections(
927 config: &InjectionConfig,
928 text: &BufferSnapshot,
929 node: Node,
930 language_registry: &LanguageRegistry,
931 depth: usize,
932 query_ranges: &[Range<usize>],
933 queue: &mut BinaryHeap<ReparseStep>,
934) -> bool {
935 let mut result = false;
936 let mut query_cursor = QueryCursorHandle::new();
937 let mut prev_match = None;
938 for query_range in query_ranges {
939 query_cursor.set_byte_range(query_range.start.saturating_sub(1)..query_range.end);
940 for mat in query_cursor.matches(&config.query, node, TextProvider(text.as_rope())) {
941 let content_ranges = mat
942 .nodes_for_capture_index(config.content_capture_ix)
943 .map(|node| node.range())
944 .collect::<Vec<_>>();
945 if content_ranges.is_empty() {
946 continue;
947 }
948
949 // Avoid duplicate matches if two changed ranges intersect the same injection.
950 let content_range =
951 content_ranges.first().unwrap().start_byte..content_ranges.last().unwrap().end_byte;
952 if let Some((last_pattern_ix, last_range)) = &prev_match {
953 if mat.pattern_index == *last_pattern_ix && content_range == *last_range {
954 continue;
955 }
956 }
957 prev_match = Some((mat.pattern_index, content_range.clone()));
958
959 let language_name = config.languages_by_pattern_ix[mat.pattern_index]
960 .as_ref()
961 .map(|s| Cow::Borrowed(s.as_ref()))
962 .or_else(|| {
963 let ix = config.language_capture_ix?;
964 let node = mat.nodes_for_capture_index(ix).next()?;
965 Some(Cow::Owned(text.text_for_range(node.byte_range()).collect()))
966 });
967
968 if let Some(language_name) = language_name {
969 if let Some(language) = language_registry.get_language(language_name.as_ref()) {
970 result = true;
971 let range = text.anchor_before(content_range.start)
972 ..text.anchor_after(content_range.end);
973 queue.push(ReparseStep {
974 depth,
975 language,
976 ranges: content_ranges,
977 range,
978 })
979 }
980 }
981 }
982 }
983 result
984}
985
986impl std::ops::Deref for SyntaxMap {
987 type Target = SyntaxSnapshot;
988
989 fn deref(&self) -> &Self::Target {
990 &self.snapshot
991 }
992}
993
994impl PartialEq for ReparseStep {
995 fn eq(&self, _: &Self) -> bool {
996 false
997 }
998}
999
1000impl Eq for ReparseStep {}
1001
1002impl PartialOrd for ReparseStep {
1003 fn partial_cmp(&self, other: &Self) -> Option<Ordering> {
1004 Some(self.cmp(&other))
1005 }
1006}
1007
1008impl Ord for ReparseStep {
1009 fn cmp(&self, other: &Self) -> Ordering {
1010 let range_a = self.range();
1011 let range_b = other.range();
1012 Ord::cmp(&other.depth, &self.depth)
1013 .then_with(|| Ord::cmp(&range_b.start, &range_a.start))
1014 .then_with(|| Ord::cmp(&range_a.end, &range_b.end))
1015 }
1016}
1017
1018impl ReparseStep {
1019 fn range(&self) -> Range<usize> {
1020 let start = self.ranges.first().map_or(0, |r| r.start_byte);
1021 let end = self.ranges.last().map_or(0, |r| r.end_byte);
1022 start..end
1023 }
1024}
1025
1026impl ChangedRegion {
1027 fn cmp(&self, other: &Self, buffer: &BufferSnapshot) -> Ordering {
1028 let range_a = &self.range;
1029 let range_b = &other.range;
1030 Ord::cmp(&self.depth, &other.depth)
1031 .then_with(|| range_a.start.cmp(&range_b.start, buffer))
1032 .then_with(|| range_b.end.cmp(&range_a.end, buffer))
1033 }
1034}
1035
1036impl ChangeRegionSet {
1037 fn start_position(&self) -> DepthAndMaxPosition {
1038 self.0
1039 .first()
1040 .map_or(DepthAndMaxPosition(usize::MAX, Anchor::MAX), |region| {
1041 DepthAndMaxPosition(region.depth, region.range.start)
1042 })
1043 }
1044
1045 fn intersects(&self, layer: &SyntaxLayer, text: &BufferSnapshot) -> bool {
1046 for region in &self.0 {
1047 if region.depth < layer.depth {
1048 continue;
1049 }
1050 if region.depth > layer.depth {
1051 break;
1052 }
1053 if region.range.end.cmp(&layer.range.start, text).is_le() {
1054 continue;
1055 }
1056 if region.range.start.cmp(&layer.range.end, text).is_ge() {
1057 break;
1058 }
1059 return true;
1060 }
1061 false
1062 }
1063
1064 fn insert(&mut self, region: ChangedRegion, text: &BufferSnapshot) {
1065 if let Err(ix) = self.0.binary_search_by(|probe| probe.cmp(®ion, text)) {
1066 self.0.insert(ix, region);
1067 }
1068 }
1069
1070 fn prune(&mut self, summary: SyntaxLayerSummary, text: &BufferSnapshot) -> bool {
1071 let prev_len = self.0.len();
1072 self.0.retain(|region| {
1073 region.depth > summary.max_depth
1074 || (region.depth == summary.max_depth
1075 && region
1076 .range
1077 .end
1078 .cmp(&summary.last_layer_range.start, text)
1079 .is_gt())
1080 });
1081 self.0.len() < prev_len
1082 }
1083}
1084
1085impl Default for SyntaxLayerSummary {
1086 fn default() -> Self {
1087 Self {
1088 max_depth: 0,
1089 min_depth: 0,
1090 range: Anchor::MAX..Anchor::MIN,
1091 last_layer_range: Anchor::MIN..Anchor::MAX,
1092 }
1093 }
1094}
1095
1096impl sum_tree::Summary for SyntaxLayerSummary {
1097 type Context = BufferSnapshot;
1098
1099 fn add_summary(&mut self, other: &Self, buffer: &Self::Context) {
1100 if other.max_depth > self.max_depth {
1101 self.max_depth = other.max_depth;
1102 self.range = other.range.clone();
1103 } else {
1104 if other.range.start.cmp(&self.range.start, buffer).is_lt() {
1105 self.range.start = other.range.start;
1106 }
1107 if other.range.end.cmp(&self.range.end, buffer).is_gt() {
1108 self.range.end = other.range.end;
1109 }
1110 }
1111 self.last_layer_range = other.last_layer_range.clone();
1112 }
1113}
1114
1115impl<'a> SeekTarget<'a, SyntaxLayerSummary, SyntaxLayerSummary> for DepthAndRange {
1116 fn cmp(&self, cursor_location: &SyntaxLayerSummary, buffer: &BufferSnapshot) -> Ordering {
1117 Ord::cmp(&self.0, &cursor_location.max_depth)
1118 .then_with(|| {
1119 self.1
1120 .start
1121 .cmp(&cursor_location.last_layer_range.start, buffer)
1122 })
1123 .then_with(|| {
1124 cursor_location
1125 .last_layer_range
1126 .end
1127 .cmp(&self.1.end, buffer)
1128 })
1129 }
1130}
1131
1132impl<'a> SeekTarget<'a, SyntaxLayerSummary, SyntaxLayerSummary> for DepthAndMaxPosition {
1133 fn cmp(&self, cursor_location: &SyntaxLayerSummary, text: &BufferSnapshot) -> Ordering {
1134 Ord::cmp(&self.0, &cursor_location.max_depth)
1135 .then_with(|| self.1.cmp(&cursor_location.range.end, text))
1136 }
1137}
1138
1139impl<'a> SeekTarget<'a, SyntaxLayerSummary, SyntaxLayerSummary> for DepthAndRangeOrMaxPosition {
1140 fn cmp(&self, cursor_location: &SyntaxLayerSummary, buffer: &BufferSnapshot) -> Ordering {
1141 if self.1.cmp(cursor_location, buffer).is_le() {
1142 return Ordering::Less;
1143 } else {
1144 self.0.cmp(cursor_location, buffer)
1145 }
1146 }
1147}
1148
1149impl sum_tree::Item for SyntaxLayer {
1150 type Summary = SyntaxLayerSummary;
1151
1152 fn summary(&self) -> Self::Summary {
1153 SyntaxLayerSummary {
1154 min_depth: self.depth,
1155 max_depth: self.depth,
1156 range: self.range.clone(),
1157 last_layer_range: self.range.clone(),
1158 }
1159 }
1160}
1161
1162impl std::fmt::Debug for SyntaxLayer {
1163 fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
1164 f.debug_struct("SyntaxLayer")
1165 .field("depth", &self.depth)
1166 .field("range", &self.range)
1167 .field("tree", &self.tree)
1168 .finish()
1169 }
1170}
1171
1172impl<'a> tree_sitter::TextProvider<'a> for TextProvider<'a> {
1173 type I = ByteChunks<'a>;
1174
1175 fn text(&mut self, node: tree_sitter::Node) -> Self::I {
1176 ByteChunks(self.0.chunks_in_range(node.byte_range()))
1177 }
1178}
1179
1180impl<'a> Iterator for ByteChunks<'a> {
1181 type Item = &'a [u8];
1182
1183 fn next(&mut self) -> Option<Self::Item> {
1184 self.0.next().map(str::as_bytes)
1185 }
1186}
1187
1188impl QueryCursorHandle {
1189 pub(crate) fn new() -> Self {
1190 let mut cursor = QUERY_CURSORS.lock().pop().unwrap_or_else(QueryCursor::new);
1191 cursor.set_match_limit(64);
1192 QueryCursorHandle(Some(cursor))
1193 }
1194}
1195
1196impl Deref for QueryCursorHandle {
1197 type Target = QueryCursor;
1198
1199 fn deref(&self) -> &Self::Target {
1200 self.0.as_ref().unwrap()
1201 }
1202}
1203
1204impl DerefMut for QueryCursorHandle {
1205 fn deref_mut(&mut self) -> &mut Self::Target {
1206 self.0.as_mut().unwrap()
1207 }
1208}
1209
1210impl Drop for QueryCursorHandle {
1211 fn drop(&mut self) {
1212 let mut cursor = self.0.take().unwrap();
1213 cursor.set_byte_range(0..usize::MAX);
1214 cursor.set_point_range(Point::zero().to_ts_point()..Point::MAX.to_ts_point());
1215 QUERY_CURSORS.lock().push(cursor)
1216 }
1217}
1218
1219pub(crate) trait ToTreeSitterPoint {
1220 fn to_ts_point(self) -> tree_sitter::Point;
1221 fn from_ts_point(point: tree_sitter::Point) -> Self;
1222}
1223
1224impl ToTreeSitterPoint for Point {
1225 fn to_ts_point(self) -> tree_sitter::Point {
1226 tree_sitter::Point::new(self.row as usize, self.column as usize)
1227 }
1228
1229 fn from_ts_point(point: tree_sitter::Point) -> Self {
1230 Point::new(point.row as u32, point.column as u32)
1231 }
1232}
1233
1234#[cfg(test)]
1235mod tests {
1236 use super::*;
1237 use crate::LanguageConfig;
1238 use rand::rngs::StdRng;
1239 use std::env;
1240 use text::{Buffer, Point};
1241 use unindent::Unindent as _;
1242 use util::test::marked_text_ranges;
1243
1244 #[gpui::test]
1245 fn test_syntax_map_layers_for_range() {
1246 let registry = Arc::new(LanguageRegistry::test());
1247 let language = Arc::new(rust_lang());
1248 registry.add(language.clone());
1249
1250 let mut buffer = Buffer::new(
1251 0,
1252 0,
1253 r#"
1254 fn a() {
1255 assert_eq!(
1256 b(vec![C {}]),
1257 vec![d.e],
1258 );
1259 println!("{}", f(|_| true));
1260 }
1261 "#
1262 .unindent(),
1263 );
1264
1265 let mut syntax_map = SyntaxMap::new();
1266 syntax_map.set_language_registry(registry.clone());
1267 syntax_map.reparse(language.clone(), &buffer);
1268
1269 assert_layers_for_range(
1270 &syntax_map,
1271 &buffer,
1272 Point::new(2, 0)..Point::new(2, 0),
1273 &[
1274 "...(function_item ... (block (expression_statement (macro_invocation...",
1275 "...(tuple_expression (call_expression ... arguments: (arguments (macro_invocation...",
1276 ],
1277 );
1278 assert_layers_for_range(
1279 &syntax_map,
1280 &buffer,
1281 Point::new(2, 14)..Point::new(2, 16),
1282 &[
1283 "...(function_item ...",
1284 "...(tuple_expression (call_expression ... arguments: (arguments (macro_invocation...",
1285 "...(array_expression (struct_expression ...",
1286 ],
1287 );
1288 assert_layers_for_range(
1289 &syntax_map,
1290 &buffer,
1291 Point::new(3, 14)..Point::new(3, 16),
1292 &[
1293 "...(function_item ...",
1294 "...(tuple_expression (call_expression ... arguments: (arguments (macro_invocation...",
1295 "...(array_expression (field_expression ...",
1296 ],
1297 );
1298 assert_layers_for_range(
1299 &syntax_map,
1300 &buffer,
1301 Point::new(5, 12)..Point::new(5, 16),
1302 &[
1303 "...(function_item ...",
1304 "...(call_expression ... (arguments (closure_expression ...",
1305 ],
1306 );
1307
1308 // Replace a vec! macro invocation with a plain slice, removing a syntactic layer.
1309 let macro_name_range = range_for_text(&buffer, "vec!");
1310 buffer.edit([(macro_name_range, "&")]);
1311 syntax_map.interpolate(&buffer);
1312 syntax_map.reparse(language.clone(), &buffer);
1313
1314 assert_layers_for_range(
1315 &syntax_map,
1316 &buffer,
1317 Point::new(2, 14)..Point::new(2, 16),
1318 &[
1319 "...(function_item ...",
1320 "...(tuple_expression (call_expression ... arguments: (arguments (reference_expression value: (array_expression...",
1321 ],
1322 );
1323
1324 // Put the vec! macro back, adding back the syntactic layer.
1325 buffer.undo();
1326 syntax_map.interpolate(&buffer);
1327 syntax_map.reparse(language.clone(), &buffer);
1328
1329 assert_layers_for_range(
1330 &syntax_map,
1331 &buffer,
1332 Point::new(2, 14)..Point::new(2, 16),
1333 &[
1334 "...(function_item ...",
1335 "...(tuple_expression (call_expression ... arguments: (arguments (macro_invocation...",
1336 "...(array_expression (struct_expression ...",
1337 ],
1338 );
1339 }
1340
1341 #[gpui::test]
1342 fn test_typing_multiple_new_injections() {
1343 let (buffer, syntax_map) = test_edit_sequence(&[
1344 "fn a() { dbg }",
1345 "fn a() { dbg«!» }",
1346 "fn a() { dbg!«()» }",
1347 "fn a() { dbg!(«b») }",
1348 "fn a() { dbg!(b«.») }",
1349 "fn a() { dbg!(b.«c») }",
1350 "fn a() { dbg!(b.c«()») }",
1351 "fn a() { dbg!(b.c(«vec»)) }",
1352 "fn a() { dbg!(b.c(vec«!»)) }",
1353 "fn a() { dbg!(b.c(vec!«[]»)) }",
1354 "fn a() { dbg!(b.c(vec![«d»])) }",
1355 "fn a() { dbg!(b.c(vec![d«.»])) }",
1356 "fn a() { dbg!(b.c(vec![d.«e»])) }",
1357 ]);
1358
1359 assert_capture_ranges(
1360 &syntax_map,
1361 &buffer,
1362 &["field"],
1363 "fn a() { dbg!(b.«c»(vec![d.«e»])) }",
1364 );
1365 }
1366
1367 #[gpui::test]
1368 fn test_pasting_new_injection_line_between_others() {
1369 let (buffer, syntax_map) = test_edit_sequence(&[
1370 "
1371 fn a() {
1372 b!(B {});
1373 c!(C {});
1374 d!(D {});
1375 e!(E {});
1376 f!(F {});
1377 g!(G {});
1378 }
1379 ",
1380 "
1381 fn a() {
1382 b!(B {});
1383 c!(C {});
1384 d!(D {});
1385 « h!(H {});
1386 » e!(E {});
1387 f!(F {});
1388 g!(G {});
1389 }
1390 ",
1391 ]);
1392
1393 assert_capture_ranges(
1394 &syntax_map,
1395 &buffer,
1396 &["struct"],
1397 "
1398 fn a() {
1399 b!(«B {}»);
1400 c!(«C {}»);
1401 d!(«D {}»);
1402 h!(«H {}»);
1403 e!(«E {}»);
1404 f!(«F {}»);
1405 g!(«G {}»);
1406 }
1407 ",
1408 );
1409 }
1410
1411 #[gpui::test]
1412 fn test_joining_injections_with_child_injections() {
1413 let (buffer, syntax_map) = test_edit_sequence(&[
1414 "
1415 fn a() {
1416 b!(
1417 c![one.two.three],
1418 d![four.five.six],
1419 );
1420 e!(
1421 f![seven.eight],
1422 );
1423 }
1424 ",
1425 "
1426 fn a() {
1427 b!(
1428 c![one.two.three],
1429 d![four.five.six],
1430 ˇ f![seven.eight],
1431 );
1432 }
1433 ",
1434 ]);
1435
1436 assert_capture_ranges(
1437 &syntax_map,
1438 &buffer,
1439 &["field"],
1440 "
1441 fn a() {
1442 b!(
1443 c![one.«two».«three»],
1444 d![four.«five».«six»],
1445 f![seven.«eight»],
1446 );
1447 }
1448 ",
1449 );
1450 }
1451
1452 #[gpui::test]
1453 fn test_editing_edges_of_injection() {
1454 test_edit_sequence(&[
1455 "
1456 fn a() {
1457 b!(c!())
1458 }
1459 ",
1460 "
1461 fn a() {
1462 «d»!(c!())
1463 }
1464 ",
1465 "
1466 fn a() {
1467 «e»d!(c!())
1468 }
1469 ",
1470 "
1471 fn a() {
1472 ed!«[»c!()«]»
1473 }
1474 ",
1475 ]);
1476 }
1477
1478 #[gpui::test]
1479 fn test_edits_preceding_and_intersecting_injection() {
1480 test_edit_sequence(&[
1481 //
1482 "const aaaaaaaaaaaa: B = c!(d(e.f));",
1483 "const aˇa: B = c!(d(eˇ));",
1484 ]);
1485 }
1486
1487 #[gpui::test]
1488 fn test_non_local_changes_create_injections() {
1489 test_edit_sequence(&[
1490 "
1491 // a! {
1492 static B: C = d;
1493 // }
1494 ",
1495 "
1496 ˇa! {
1497 static B: C = d;
1498 ˇ}
1499 ",
1500 ]);
1501 }
1502
1503 #[gpui::test]
1504 fn test_creating_many_injections_in_one_edit() {
1505 test_edit_sequence(&[
1506 "
1507 fn a() {
1508 one(Two::three(3));
1509 four(Five::six(6));
1510 seven(Eight::nine(9));
1511 }
1512 ",
1513 "
1514 fn a() {
1515 one«!»(Two::three(3));
1516 four«!»(Five::six(6));
1517 seven«!»(Eight::nine(9));
1518 }
1519 ",
1520 "
1521 fn a() {
1522 one!(Two::three«!»(3));
1523 four!(Five::six«!»(6));
1524 seven!(Eight::nine«!»(9));
1525 }
1526 ",
1527 ]);
1528 }
1529
1530 #[gpui::test]
1531 fn test_editing_across_injection_boundary() {
1532 test_edit_sequence(&[
1533 "
1534 fn one() {
1535 two();
1536 three!(
1537 three.four,
1538 five.six,
1539 );
1540 }
1541 ",
1542 "
1543 fn one() {
1544 two();
1545 th«irty_five![»
1546 three.four,
1547 five.six,
1548 « seven.eight,
1549 ];»
1550 }
1551 ",
1552 ]);
1553 }
1554
1555 #[gpui::test]
1556 fn test_removing_injection_by_replacing_across_boundary() {
1557 test_edit_sequence(&[
1558 "
1559 fn one() {
1560 two!(
1561 three.four,
1562 );
1563 }
1564 ",
1565 "
1566 fn one() {
1567 t«en
1568 .eleven(
1569 twelve,
1570 »
1571 three.four,
1572 );
1573 }
1574 ",
1575 ]);
1576 }
1577
1578 #[gpui::test(iterations = 100)]
1579 fn test_random_syntax_map_edits(mut rng: StdRng) {
1580 let operations = env::var("OPERATIONS")
1581 .map(|i| i.parse().expect("invalid `OPERATIONS` variable"))
1582 .unwrap_or(10);
1583
1584 let text = r#"
1585 fn test_something() {
1586 let vec = vec![5, 1, 3, 8];
1587 assert_eq!(
1588 vec
1589 .into_iter()
1590 .map(|i| i * 2)
1591 .collect::<Vec<usize>>(),
1592 vec![
1593 5 * 2, 1 * 2, 3 * 2, 8 * 2
1594 ],
1595 );
1596 }
1597 "#
1598 .unindent()
1599 .repeat(2);
1600
1601 let registry = Arc::new(LanguageRegistry::test());
1602 let language = Arc::new(rust_lang());
1603 registry.add(language.clone());
1604 let mut buffer = Buffer::new(0, 0, text);
1605
1606 let mut syntax_map = SyntaxMap::new();
1607 syntax_map.set_language_registry(registry.clone());
1608 syntax_map.reparse(language.clone(), &buffer);
1609
1610 let mut reference_syntax_map = SyntaxMap::new();
1611 reference_syntax_map.set_language_registry(registry.clone());
1612
1613 log::info!("initial text:\n{}", buffer.text());
1614
1615 for _ in 0..operations {
1616 let prev_buffer = buffer.snapshot();
1617 let prev_syntax_map = syntax_map.snapshot();
1618
1619 buffer.randomly_edit(&mut rng, 3);
1620 log::info!("text:\n{}", buffer.text());
1621
1622 syntax_map.interpolate(&buffer);
1623 check_interpolation(&prev_syntax_map, &syntax_map, &prev_buffer, &buffer);
1624
1625 syntax_map.reparse(language.clone(), &buffer);
1626
1627 reference_syntax_map.clear();
1628 reference_syntax_map.reparse(language.clone(), &buffer);
1629 }
1630
1631 for i in 0..operations {
1632 let i = operations - i - 1;
1633 buffer.undo();
1634 log::info!("undoing operation {}", i);
1635 log::info!("text:\n{}", buffer.text());
1636
1637 syntax_map.interpolate(&buffer);
1638 syntax_map.reparse(language.clone(), &buffer);
1639
1640 reference_syntax_map.clear();
1641 reference_syntax_map.reparse(language.clone(), &buffer);
1642 assert_eq!(
1643 syntax_map.layers(&buffer).len(),
1644 reference_syntax_map.layers(&buffer).len(),
1645 "wrong number of layers after undoing edit {i}"
1646 );
1647 }
1648
1649 let layers = syntax_map.layers(&buffer);
1650 let reference_layers = reference_syntax_map.layers(&buffer);
1651 for (edited_layer, reference_layer) in layers.into_iter().zip(reference_layers.into_iter())
1652 {
1653 assert_eq!(edited_layer.node.to_sexp(), reference_layer.node.to_sexp());
1654 assert_eq!(edited_layer.node.range(), reference_layer.node.range());
1655 }
1656 }
1657
1658 fn check_interpolation(
1659 old_syntax_map: &SyntaxSnapshot,
1660 new_syntax_map: &SyntaxSnapshot,
1661 old_buffer: &BufferSnapshot,
1662 new_buffer: &BufferSnapshot,
1663 ) {
1664 let edits = new_buffer
1665 .edits_since::<usize>(&old_buffer.version())
1666 .collect::<Vec<_>>();
1667
1668 for (old_layer, new_layer) in old_syntax_map
1669 .layers
1670 .iter()
1671 .zip(new_syntax_map.layers.iter())
1672 {
1673 assert_eq!(old_layer.range, new_layer.range);
1674 let old_start_byte = old_layer.range.start.to_offset(old_buffer);
1675 let new_start_byte = new_layer.range.start.to_offset(new_buffer);
1676 let old_start_point = old_layer.range.start.to_point(old_buffer).to_ts_point();
1677 let new_start_point = new_layer.range.start.to_point(new_buffer).to_ts_point();
1678 let old_node = old_layer
1679 .tree
1680 .root_node_with_offset(old_start_byte, old_start_point);
1681 let new_node = new_layer
1682 .tree
1683 .root_node_with_offset(new_start_byte, new_start_point);
1684 check_node_edits(
1685 old_layer.depth,
1686 &old_layer.range,
1687 old_node,
1688 new_node,
1689 old_buffer,
1690 new_buffer,
1691 &edits,
1692 );
1693 }
1694
1695 fn check_node_edits(
1696 depth: usize,
1697 range: &Range<Anchor>,
1698 old_node: Node,
1699 new_node: Node,
1700 old_buffer: &BufferSnapshot,
1701 new_buffer: &BufferSnapshot,
1702 edits: &[text::Edit<usize>],
1703 ) {
1704 assert_eq!(old_node.kind(), new_node.kind());
1705
1706 let old_range = old_node.byte_range();
1707 let new_range = new_node.byte_range();
1708
1709 let is_edited = edits
1710 .iter()
1711 .any(|edit| edit.new.start < new_range.end && edit.new.end > new_range.start);
1712 if is_edited {
1713 assert!(
1714 new_node.has_changes(),
1715 concat!(
1716 "failed to mark node as edited.\n",
1717 "layer depth: {}, old layer range: {:?}, new layer range: {:?},\n",
1718 "node kind: {}, old node range: {:?}, new node range: {:?}",
1719 ),
1720 depth,
1721 range.to_offset(old_buffer),
1722 range.to_offset(new_buffer),
1723 new_node.kind(),
1724 old_range,
1725 new_range,
1726 );
1727 }
1728
1729 if !new_node.has_changes() {
1730 assert_eq!(
1731 old_buffer
1732 .text_for_range(old_range.clone())
1733 .collect::<String>(),
1734 new_buffer
1735 .text_for_range(new_range.clone())
1736 .collect::<String>(),
1737 concat!(
1738 "mismatched text for node\n",
1739 "layer depth: {}, old layer range: {:?}, new layer range: {:?},\n",
1740 "node kind: {}, old node range:{:?}, new node range:{:?}",
1741 ),
1742 depth,
1743 range.to_offset(old_buffer),
1744 range.to_offset(new_buffer),
1745 new_node.kind(),
1746 old_range,
1747 new_range,
1748 );
1749 }
1750
1751 for i in 0..new_node.child_count() {
1752 check_node_edits(
1753 depth,
1754 range,
1755 old_node.child(i).unwrap(),
1756 new_node.child(i).unwrap(),
1757 old_buffer,
1758 new_buffer,
1759 edits,
1760 )
1761 }
1762 }
1763 }
1764
1765 fn test_edit_sequence(steps: &[&str]) -> (Buffer, SyntaxMap) {
1766 let registry = Arc::new(LanguageRegistry::test());
1767 let language = Arc::new(rust_lang());
1768 registry.add(language.clone());
1769 let mut buffer = Buffer::new(0, 0, Default::default());
1770
1771 let mut mutated_syntax_map = SyntaxMap::new();
1772 mutated_syntax_map.set_language_registry(registry.clone());
1773 mutated_syntax_map.reparse(language.clone(), &buffer);
1774
1775 for (i, marked_string) in steps.into_iter().enumerate() {
1776 edit_buffer(&mut buffer, &marked_string.unindent());
1777
1778 // Reparse the syntax map
1779 mutated_syntax_map.interpolate(&buffer);
1780 mutated_syntax_map.reparse(language.clone(), &buffer);
1781
1782 // Create a second syntax map from scratch
1783 let mut reference_syntax_map = SyntaxMap::new();
1784 reference_syntax_map.set_language_registry(registry.clone());
1785 reference_syntax_map.reparse(language.clone(), &buffer);
1786
1787 // Compare the mutated syntax map to the new syntax map
1788 let mutated_layers = mutated_syntax_map.layers(&buffer);
1789 let reference_layers = reference_syntax_map.layers(&buffer);
1790 assert_eq!(
1791 mutated_layers.len(),
1792 reference_layers.len(),
1793 "wrong number of layers at step {i}"
1794 );
1795 for (edited_layer, reference_layer) in
1796 mutated_layers.into_iter().zip(reference_layers.into_iter())
1797 {
1798 assert_eq!(
1799 edited_layer.node.to_sexp(),
1800 reference_layer.node.to_sexp(),
1801 "different layer at step {i}"
1802 );
1803 assert_eq!(
1804 edited_layer.node.range(),
1805 reference_layer.node.range(),
1806 "different layer at step {i}"
1807 );
1808 }
1809 }
1810
1811 (buffer, mutated_syntax_map)
1812 }
1813
1814 fn rust_lang() -> Language {
1815 Language::new(
1816 LanguageConfig {
1817 name: "Rust".into(),
1818 path_suffixes: vec!["rs".to_string()],
1819 ..Default::default()
1820 },
1821 Some(tree_sitter_rust::language()),
1822 )
1823 .with_highlights_query(
1824 r#"
1825 (field_identifier) @field
1826 (struct_expression) @struct
1827 "#,
1828 )
1829 .unwrap()
1830 .with_injection_query(
1831 r#"
1832 (macro_invocation
1833 (token_tree) @content
1834 (#set! "language" "rust"))
1835 "#,
1836 )
1837 .unwrap()
1838 }
1839
1840 fn range_for_text(buffer: &Buffer, text: &str) -> Range<usize> {
1841 let start = buffer.as_rope().to_string().find(text).unwrap();
1842 start..start + text.len()
1843 }
1844
1845 fn assert_layers_for_range(
1846 syntax_map: &SyntaxMap,
1847 buffer: &BufferSnapshot,
1848 range: Range<Point>,
1849 expected_layers: &[&str],
1850 ) {
1851 let layers = syntax_map.layers_for_range(range, &buffer);
1852 assert_eq!(
1853 layers.len(),
1854 expected_layers.len(),
1855 "wrong number of layers"
1856 );
1857 for (i, (SyntaxLayerInfo { node, .. }, expected_s_exp)) in
1858 layers.iter().zip(expected_layers.iter()).enumerate()
1859 {
1860 let actual_s_exp = node.to_sexp();
1861 assert!(
1862 string_contains_sequence(
1863 &actual_s_exp,
1864 &expected_s_exp.split("...").collect::<Vec<_>>()
1865 ),
1866 "layer {i}:\n\nexpected: {expected_s_exp}\nactual: {actual_s_exp}",
1867 );
1868 }
1869 }
1870
1871 fn assert_capture_ranges(
1872 syntax_map: &SyntaxMap,
1873 buffer: &BufferSnapshot,
1874 highlight_query_capture_names: &[&str],
1875 marked_string: &str,
1876 ) {
1877 let mut actual_ranges = Vec::<Range<usize>>::new();
1878 let captures = syntax_map.captures(0..buffer.len(), buffer, |grammar| {
1879 grammar.highlights_query.as_ref()
1880 });
1881 let queries = captures
1882 .grammars()
1883 .iter()
1884 .map(|grammar| grammar.highlights_query.as_ref().unwrap())
1885 .collect::<Vec<_>>();
1886 for capture in captures {
1887 let name = &queries[capture.grammar_index].capture_names()[capture.index as usize];
1888 if highlight_query_capture_names.contains(&name.as_str()) {
1889 actual_ranges.push(capture.node.byte_range());
1890 }
1891 }
1892
1893 let (text, expected_ranges) = marked_text_ranges(&marked_string.unindent(), false);
1894 assert_eq!(text, buffer.text());
1895 assert_eq!(actual_ranges, expected_ranges);
1896 }
1897
1898 fn edit_buffer(buffer: &mut Buffer, marked_string: &str) {
1899 let old_text = buffer.text();
1900 let (new_text, mut ranges) = marked_text_ranges(marked_string, false);
1901 if ranges.is_empty() {
1902 ranges.push(0..new_text.len());
1903 }
1904
1905 assert_eq!(
1906 old_text[..ranges[0].start],
1907 new_text[..ranges[0].start],
1908 "invalid edit"
1909 );
1910
1911 let mut delta = 0;
1912 let mut edits = Vec::new();
1913 let mut ranges = ranges.into_iter().peekable();
1914
1915 while let Some(inserted_range) = ranges.next() {
1916 let new_start = inserted_range.start;
1917 let old_start = (new_start as isize - delta) as usize;
1918
1919 let following_text = if let Some(next_range) = ranges.peek() {
1920 &new_text[inserted_range.end..next_range.start]
1921 } else {
1922 &new_text[inserted_range.end..]
1923 };
1924
1925 let inserted_len = inserted_range.len();
1926 let deleted_len = old_text[old_start..]
1927 .find(following_text)
1928 .expect("invalid edit");
1929
1930 let old_range = old_start..old_start + deleted_len;
1931 edits.push((old_range, new_text[inserted_range].to_string()));
1932 delta += inserted_len as isize - deleted_len as isize;
1933 }
1934
1935 assert_eq!(
1936 old_text.len() as isize + delta,
1937 new_text.len() as isize,
1938 "invalid edit"
1939 );
1940
1941 buffer.edit(edits);
1942 }
1943
1944 pub fn string_contains_sequence(text: &str, parts: &[&str]) -> bool {
1945 let mut last_part_end = 0;
1946 for part in parts {
1947 if let Some(start_ix) = text[last_part_end..].find(part) {
1948 last_part_end = start_ix + part.len();
1949 } else {
1950 return false;
1951 }
1952 }
1953 true
1954 }
1955}