This function (and the similar `ts_tree_cursor_goto_first_child_for_byte`) allows you to efficiently seek the tree cursor to a given position, exploiting the tree's internal balancing, without having to visit all of the preceding siblings of each node.
2085 lines
72 KiB
Rust
2085 lines
72 KiB
Rust
mod ffi;
|
|
mod util;
|
|
|
|
#[cfg(feature = "allocation-tracking")]
|
|
pub mod allocations;
|
|
|
|
#[cfg(unix)]
|
|
use std::os::unix::io::AsRawFd;
|
|
|
|
use std::ffi::CStr;
|
|
use std::marker::PhantomData;
|
|
use std::mem::MaybeUninit;
|
|
use std::os::raw::{c_char, c_void};
|
|
use std::ptr::NonNull;
|
|
use std::sync::atomic::AtomicUsize;
|
|
use std::{char, error, fmt, hash, iter, ptr, slice, str, u16};
|
|
|
|
/// The latest ABI version that is supported by the current version of the
|
|
/// library.
|
|
///
|
|
/// When Languages are generated by the Tree-sitter CLI, they are
|
|
/// assigned an ABI version number that corresponds to the current CLI version.
|
|
/// The Tree-sitter library is generally backwards-compatible with languages
|
|
/// generated using older CLI versions, but is not forwards-compatible.
|
|
pub const LANGUAGE_VERSION: usize = ffi::TREE_SITTER_LANGUAGE_VERSION;
|
|
|
|
/// The earliest ABI version that is supported by the current version of the
|
|
/// library.
|
|
pub const MIN_COMPATIBLE_LANGUAGE_VERSION: usize = ffi::TREE_SITTER_MIN_COMPATIBLE_LANGUAGE_VERSION;
|
|
|
|
pub const PARSER_HEADER: &'static str = include_str!("../include/tree_sitter/parser.h");
|
|
|
|
/// An opaque object that defines how to parse a particular language. The code for each
|
|
/// `Language` is generated by the Tree-sitter CLI.
|
|
#[derive(Clone, Copy, Debug, PartialEq, Eq)]
|
|
#[repr(transparent)]
|
|
pub struct Language(*const ffi::TSLanguage);
|
|
|
|
/// A tree that represents the syntactic structure of a source code file.
|
|
pub struct Tree(NonNull<ffi::TSTree>);
|
|
|
|
/// A position in a multi-line text document, in terms of rows and columns.
|
|
///
|
|
/// Rows and columns are zero-based.
|
|
#[derive(Clone, Copy, Debug, Default, PartialEq, Eq, Hash, PartialOrd, Ord)]
|
|
pub struct Point {
|
|
pub row: usize,
|
|
pub column: usize,
|
|
}
|
|
|
|
/// A range of positions in a multi-line text document, both in terms of bytes and of
|
|
/// rows and columns.
|
|
#[derive(Clone, Copy, Debug, PartialEq, Eq, Hash, PartialOrd, Ord)]
|
|
pub struct Range {
|
|
pub start_byte: usize,
|
|
pub end_byte: usize,
|
|
pub start_point: Point,
|
|
pub end_point: Point,
|
|
}
|
|
|
|
/// A summary of a change to a text document.
|
|
#[derive(Clone, Copy, Debug, PartialEq, Eq)]
|
|
pub struct InputEdit {
|
|
pub start_byte: usize,
|
|
pub old_end_byte: usize,
|
|
pub new_end_byte: usize,
|
|
pub start_position: Point,
|
|
pub old_end_position: Point,
|
|
pub new_end_position: Point,
|
|
}
|
|
|
|
/// A single node within a syntax `Tree`.
|
|
#[derive(Clone, Copy)]
|
|
#[repr(transparent)]
|
|
pub struct Node<'a>(ffi::TSNode, PhantomData<&'a ()>);
|
|
|
|
/// A stateful object that this is used to produce a `Tree` based on some source code.
|
|
pub struct Parser(NonNull<ffi::TSParser>);
|
|
|
|
/// A type of log message.
|
|
#[derive(Debug, PartialEq, Eq)]
|
|
pub enum LogType {
|
|
Parse,
|
|
Lex,
|
|
}
|
|
|
|
/// A callback that receives log messages during parser.
|
|
type Logger<'a> = Box<dyn FnMut(LogType, &str) + 'a>;
|
|
|
|
/// A stateful object for walking a syntax `Tree` efficiently.
|
|
pub struct TreeCursor<'a>(ffi::TSTreeCursor, PhantomData<&'a ()>);
|
|
|
|
/// A set of patterns that match nodes in a syntax tree.
|
|
#[derive(Debug)]
|
|
pub struct Query {
|
|
ptr: NonNull<ffi::TSQuery>,
|
|
capture_names: Vec<String>,
|
|
text_predicates: Vec<Box<[TextPredicate]>>,
|
|
property_settings: Vec<Box<[QueryProperty]>>,
|
|
property_predicates: Vec<Box<[(QueryProperty, bool)]>>,
|
|
general_predicates: Vec<Box<[QueryPredicate]>>,
|
|
}
|
|
|
|
/// A stateful object for executing a `Query` on a syntax `Tree`.
|
|
pub struct QueryCursor {
|
|
ptr: NonNull<ffi::TSQueryCursor>,
|
|
}
|
|
|
|
/// A key-value pair associated with a particular pattern in a `Query`.
|
|
#[derive(Debug, PartialEq, Eq)]
|
|
pub struct QueryProperty {
|
|
pub key: Box<str>,
|
|
pub value: Option<Box<str>>,
|
|
pub capture_id: Option<usize>,
|
|
}
|
|
|
|
#[derive(Debug, PartialEq, Eq)]
|
|
pub enum QueryPredicateArg {
|
|
Capture(u32),
|
|
String(Box<str>),
|
|
}
|
|
|
|
/// A key-value pair associated with a particular pattern in a `Query`.
|
|
#[derive(Debug, PartialEq, Eq)]
|
|
pub struct QueryPredicate {
|
|
pub operator: Box<str>,
|
|
pub args: Vec<QueryPredicateArg>,
|
|
}
|
|
|
|
/// A match of a `Query` to a particular set of `Node`s.
|
|
pub struct QueryMatch<'cursor, 'tree> {
|
|
pub pattern_index: usize,
|
|
pub captures: &'cursor [QueryCapture<'tree>],
|
|
id: u32,
|
|
cursor: *mut ffi::TSQueryCursor,
|
|
}
|
|
|
|
/// A sequence of `QueryMatch`es associated with a given `QueryCursor`.
|
|
pub struct QueryMatches<'a, 'tree: 'a, T: TextProvider<'a>> {
|
|
ptr: *mut ffi::TSQueryCursor,
|
|
query: &'a Query,
|
|
text_provider: T,
|
|
buffer1: Vec<u8>,
|
|
buffer2: Vec<u8>,
|
|
_tree: PhantomData<&'tree ()>,
|
|
}
|
|
|
|
/// A sequence of `QueryCapture`s associated with a given `QueryCursor`.
|
|
pub struct QueryCaptures<'a, 'tree: 'a, T: TextProvider<'a>> {
|
|
ptr: *mut ffi::TSQueryCursor,
|
|
query: &'a Query,
|
|
text_provider: T,
|
|
buffer1: Vec<u8>,
|
|
buffer2: Vec<u8>,
|
|
_tree: PhantomData<&'tree ()>,
|
|
}
|
|
|
|
pub trait TextProvider<'a> {
|
|
type I: Iterator<Item = &'a [u8]> + 'a;
|
|
fn text(&mut self, node: Node) -> Self::I;
|
|
}
|
|
|
|
/// A particular `Node` that has been captured with a particular name within a `Query`.
|
|
#[derive(Clone, Copy, Debug)]
|
|
#[repr(C)]
|
|
pub struct QueryCapture<'a> {
|
|
pub node: Node<'a>,
|
|
pub index: u32,
|
|
}
|
|
|
|
/// An error that occurred when trying to assign an incompatible `Language` to a `Parser`.
|
|
#[derive(Debug, PartialEq, Eq)]
|
|
pub struct LanguageError {
|
|
version: usize,
|
|
}
|
|
|
|
/// An error that occurred in `Parser::set_included_ranges`.
|
|
#[derive(Debug, PartialEq, Eq)]
|
|
pub struct IncludedRangesError(pub usize);
|
|
|
|
/// An error that occurred when trying to create a `Query`.
|
|
#[derive(Debug, PartialEq, Eq)]
|
|
pub struct QueryError {
|
|
pub row: usize,
|
|
pub column: usize,
|
|
pub offset: usize,
|
|
pub message: String,
|
|
pub kind: QueryErrorKind,
|
|
}
|
|
|
|
#[derive(Debug, PartialEq, Eq)]
|
|
pub enum QueryErrorKind {
|
|
Syntax,
|
|
NodeType,
|
|
Field,
|
|
Capture,
|
|
Predicate,
|
|
Structure,
|
|
}
|
|
|
|
trait TextCallback<'a> {
|
|
fn call(&mut self, node: Node);
|
|
fn next_chunk(&mut self) -> Option<&'a [u8]>;
|
|
}
|
|
|
|
#[derive(Debug)]
|
|
enum TextPredicate {
|
|
CaptureEqString(u32, String, bool),
|
|
CaptureEqCapture(u32, u32, bool),
|
|
CaptureMatchString(u32, regex::bytes::Regex, bool),
|
|
}
|
|
|
|
// TODO: Remove this struct at at some point. If `core::str::lossy::Utf8Lossy`
|
|
// is ever stabilized.
|
|
pub struct LossyUtf8<'a> {
|
|
bytes: &'a [u8],
|
|
in_replacement: bool,
|
|
}
|
|
|
|
impl Language {
|
|
/// Get the ABI version number that indicates which version of the Tree-sitter CLI
|
|
/// that was used to generate this `Language`.
|
|
pub fn version(&self) -> usize {
|
|
unsafe { ffi::ts_language_version(self.0) as usize }
|
|
}
|
|
|
|
/// Get the number of distinct node types in this language.
|
|
pub fn node_kind_count(&self) -> usize {
|
|
unsafe { ffi::ts_language_symbol_count(self.0) as usize }
|
|
}
|
|
|
|
/// Get the name of the node kind for the given numerical id.
|
|
pub fn node_kind_for_id(&self, id: u16) -> Option<&'static str> {
|
|
let ptr = unsafe { ffi::ts_language_symbol_name(self.0, id) };
|
|
if ptr.is_null() {
|
|
None
|
|
} else {
|
|
Some(unsafe { CStr::from_ptr(ptr) }.to_str().unwrap())
|
|
}
|
|
}
|
|
|
|
/// Get the numeric id for the given node kind.
|
|
pub fn id_for_node_kind(&self, kind: &str, named: bool) -> u16 {
|
|
unsafe {
|
|
ffi::ts_language_symbol_for_name(
|
|
self.0,
|
|
kind.as_bytes().as_ptr() as *const c_char,
|
|
kind.len() as u32,
|
|
named,
|
|
)
|
|
}
|
|
}
|
|
|
|
/// Check if the node type for the given numerical id is named (as opposed
|
|
/// to an anonymous node type).
|
|
pub fn node_kind_is_named(&self, id: u16) -> bool {
|
|
unsafe { ffi::ts_language_symbol_type(self.0, id) == ffi::TSSymbolType_TSSymbolTypeRegular }
|
|
}
|
|
|
|
pub fn node_kind_is_visible(&self, id: u16) -> bool {
|
|
unsafe {
|
|
ffi::ts_language_symbol_type(self.0, id) <= ffi::TSSymbolType_TSSymbolTypeAnonymous
|
|
}
|
|
}
|
|
|
|
/// Get the number of distinct field names in this language.
|
|
pub fn field_count(&self) -> usize {
|
|
unsafe { ffi::ts_language_field_count(self.0) as usize }
|
|
}
|
|
|
|
/// Get the field names for the given numerical id.
|
|
pub fn field_name_for_id(&self, field_id: u16) -> Option<&'static str> {
|
|
let ptr = unsafe { ffi::ts_language_field_name_for_id(self.0, field_id) };
|
|
if ptr.is_null() {
|
|
None
|
|
} else {
|
|
Some(unsafe { CStr::from_ptr(ptr) }.to_str().unwrap())
|
|
}
|
|
}
|
|
|
|
/// Get the numerical id for the given field name.
|
|
pub fn field_id_for_name(&self, field_name: impl AsRef<[u8]>) -> Option<u16> {
|
|
let field_name = field_name.as_ref();
|
|
let id = unsafe {
|
|
ffi::ts_language_field_id_for_name(
|
|
self.0,
|
|
field_name.as_ptr() as *const c_char,
|
|
field_name.len() as u32,
|
|
)
|
|
};
|
|
if id == 0 {
|
|
None
|
|
} else {
|
|
Some(id)
|
|
}
|
|
}
|
|
}
|
|
|
|
impl Parser {
|
|
/// Create a new parser.
|
|
pub fn new() -> Parser {
|
|
unsafe {
|
|
let parser = ffi::ts_parser_new();
|
|
Parser(NonNull::new_unchecked(parser))
|
|
}
|
|
}
|
|
|
|
/// Set the language that the parser should use for parsing.
|
|
///
|
|
/// Returns a Result indicating whether or not the language was successfully
|
|
/// assigned. True means assignment succeeded. False means there was a version
|
|
/// mismatch: the language was generated with an incompatible version of the
|
|
/// Tree-sitter CLI. Check the language's version using [Language::version]
|
|
/// and compare it to this library's [LANGUAGE_VERSION](LANGUAGE_VERSION) and
|
|
/// [MIN_COMPATIBLE_LANGUAGE_VERSION](MIN_COMPATIBLE_LANGUAGE_VERSION) constants.
|
|
pub fn set_language(&mut self, language: Language) -> Result<(), LanguageError> {
|
|
let version = language.version();
|
|
if version < MIN_COMPATIBLE_LANGUAGE_VERSION || version > LANGUAGE_VERSION {
|
|
Err(LanguageError { version })
|
|
} else {
|
|
unsafe {
|
|
ffi::ts_parser_set_language(self.0.as_ptr(), language.0);
|
|
}
|
|
Ok(())
|
|
}
|
|
}
|
|
|
|
/// Get the parser's current language.
|
|
pub fn language(&self) -> Option<Language> {
|
|
let ptr = unsafe { ffi::ts_parser_language(self.0.as_ptr()) };
|
|
if ptr.is_null() {
|
|
None
|
|
} else {
|
|
Some(Language(ptr))
|
|
}
|
|
}
|
|
|
|
/// Get the parser's current logger.
|
|
pub fn logger(&self) -> Option<&Logger> {
|
|
let logger = unsafe { ffi::ts_parser_logger(self.0.as_ptr()) };
|
|
unsafe { (logger.payload as *mut Logger).as_ref() }
|
|
}
|
|
|
|
/// Set the logging callback that a parser should use during parsing.
|
|
pub fn set_logger(&mut self, logger: Option<Logger>) {
|
|
let prev_logger = unsafe { ffi::ts_parser_logger(self.0.as_ptr()) };
|
|
if !prev_logger.payload.is_null() {
|
|
drop(unsafe { Box::from_raw(prev_logger.payload as *mut Logger) });
|
|
}
|
|
|
|
let c_logger;
|
|
if let Some(logger) = logger {
|
|
let container = Box::new(logger);
|
|
|
|
unsafe extern "C" fn log(
|
|
payload: *mut c_void,
|
|
c_log_type: ffi::TSLogType,
|
|
c_message: *const c_char,
|
|
) {
|
|
let callback = (payload as *mut Logger).as_mut().unwrap();
|
|
if let Ok(message) = CStr::from_ptr(c_message).to_str() {
|
|
let log_type = if c_log_type == ffi::TSLogType_TSLogTypeParse {
|
|
LogType::Parse
|
|
} else {
|
|
LogType::Lex
|
|
};
|
|
callback(log_type, message);
|
|
}
|
|
}
|
|
|
|
let raw_container = Box::into_raw(container);
|
|
|
|
c_logger = ffi::TSLogger {
|
|
payload: raw_container as *mut c_void,
|
|
log: Some(log),
|
|
};
|
|
} else {
|
|
c_logger = ffi::TSLogger {
|
|
payload: ptr::null_mut(),
|
|
log: None,
|
|
};
|
|
}
|
|
|
|
unsafe { ffi::ts_parser_set_logger(self.0.as_ptr(), c_logger) };
|
|
}
|
|
|
|
/// Set the destination to which the parser should write debugging graphs
|
|
/// during parsing. The graphs are formatted in the DOT language. You may want
|
|
/// to pipe these graphs directly to a `dot(1)` process in order to generate
|
|
/// SVG output.
|
|
#[cfg(unix)]
|
|
pub fn print_dot_graphs(&mut self, file: &impl AsRawFd) {
|
|
let fd = file.as_raw_fd();
|
|
unsafe { ffi::ts_parser_print_dot_graphs(self.0.as_ptr(), ffi::dup(fd)) }
|
|
}
|
|
|
|
/// Stop the parser from printing debugging graphs while parsing.
|
|
pub fn stop_printing_dot_graphs(&mut self) {
|
|
unsafe { ffi::ts_parser_print_dot_graphs(self.0.as_ptr(), -1) }
|
|
}
|
|
|
|
/// Parse a slice of UTF8 text.
|
|
///
|
|
/// # Arguments:
|
|
/// * `text` The UTF8-encoded text to parse.
|
|
/// * `old_tree` A previous syntax tree parsed from the same document.
|
|
/// If the text of the document has changed since `old_tree` was
|
|
/// created, then you must edit `old_tree` to match the new text using
|
|
/// [Tree::edit].
|
|
///
|
|
/// Returns a [Tree] if parsing succeeded, or `None` if:
|
|
/// * The parser has not yet had a language assigned with [Parser::set_language]
|
|
/// * The timeout set with [Parser::set_timeout_micros] expired
|
|
/// * The cancellation flag set with [Parser::set_cancellation_flag] was flipped
|
|
pub fn parse(&mut self, text: impl AsRef<[u8]>, old_tree: Option<&Tree>) -> Option<Tree> {
|
|
let bytes = text.as_ref();
|
|
let len = bytes.len();
|
|
self.parse_with(
|
|
&mut |i, _| if i < len { &bytes[i..] } else { &[] },
|
|
old_tree,
|
|
)
|
|
}
|
|
|
|
/// Parse a slice of UTF16 text.
|
|
///
|
|
/// # Arguments:
|
|
/// * `text` The UTF16-encoded text to parse.
|
|
/// * `old_tree` A previous syntax tree parsed from the same document.
|
|
/// If the text of the document has changed since `old_tree` was
|
|
/// created, then you must edit `old_tree` to match the new text using
|
|
/// [Tree::edit].
|
|
pub fn parse_utf16(
|
|
&mut self,
|
|
input: impl AsRef<[u16]>,
|
|
old_tree: Option<&Tree>,
|
|
) -> Option<Tree> {
|
|
let code_points = input.as_ref();
|
|
let len = code_points.len();
|
|
self.parse_utf16_with(
|
|
&mut |i, _| if i < len { &code_points[i..] } else { &[] },
|
|
old_tree,
|
|
)
|
|
}
|
|
|
|
/// Parse UTF8 text provided in chunks by a callback.
|
|
///
|
|
/// # Arguments:
|
|
/// * `callback` A function that takes a byte offset and position and
|
|
/// returns a slice of UTF8-encoded text starting at that byte offset
|
|
/// and position. The slices can be of any length. If the given position
|
|
/// is at the end of the text, the callback should return an empty slice.
|
|
/// * `old_tree` A previous syntax tree parsed from the same document.
|
|
/// If the text of the document has changed since `old_tree` was
|
|
/// created, then you must edit `old_tree` to match the new text using
|
|
/// [Tree::edit].
|
|
pub fn parse_with<'a, T: AsRef<[u8]>, F: FnMut(usize, Point) -> T>(
|
|
&mut self,
|
|
callback: &mut F,
|
|
old_tree: Option<&Tree>,
|
|
) -> Option<Tree> {
|
|
// A pointer to this payload is passed on every call to the `read` C function.
|
|
// The payload contains two things:
|
|
// 1. A reference to the rust `callback`.
|
|
// 2. The text that was returned from the previous call to `callback`.
|
|
// This allows the callback to return owned values like vectors.
|
|
let mut payload: (&mut F, Option<T>) = (callback, None);
|
|
|
|
// This C function is passed to Tree-sitter as the input callback.
|
|
unsafe extern "C" fn read<'a, T: AsRef<[u8]>, F: FnMut(usize, Point) -> T>(
|
|
payload: *mut c_void,
|
|
byte_offset: u32,
|
|
position: ffi::TSPoint,
|
|
bytes_read: *mut u32,
|
|
) -> *const c_char {
|
|
let (callback, text) = (payload as *mut (&mut F, Option<T>)).as_mut().unwrap();
|
|
*text = Some(callback(byte_offset as usize, position.into()));
|
|
let slice = text.as_ref().unwrap().as_ref();
|
|
*bytes_read = slice.len() as u32;
|
|
return slice.as_ptr() as *const c_char;
|
|
}
|
|
|
|
let c_input = ffi::TSInput {
|
|
payload: &mut payload as *mut (&mut F, Option<T>) as *mut c_void,
|
|
read: Some(read::<T, F>),
|
|
encoding: ffi::TSInputEncoding_TSInputEncodingUTF8,
|
|
};
|
|
|
|
let c_old_tree = old_tree.map_or(ptr::null_mut(), |t| t.0.as_ptr());
|
|
unsafe {
|
|
let c_new_tree = ffi::ts_parser_parse(self.0.as_ptr(), c_old_tree, c_input);
|
|
NonNull::new(c_new_tree).map(Tree)
|
|
}
|
|
}
|
|
|
|
/// Parse UTF16 text provided in chunks by a callback.
|
|
///
|
|
/// # Arguments:
|
|
/// * `callback` A function that takes a code point offset and position and
|
|
/// returns a slice of UTF16-encoded text starting at that byte offset
|
|
/// and position. The slices can be of any length. If the given position
|
|
/// is at the end of the text, the callback should return an empty slice.
|
|
/// * `old_tree` A previous syntax tree parsed from the same document.
|
|
/// If the text of the document has changed since `old_tree` was
|
|
/// created, then you must edit `old_tree` to match the new text using
|
|
/// [Tree::edit].
|
|
pub fn parse_utf16_with<'a, T: AsRef<[u16]>, F: FnMut(usize, Point) -> T>(
|
|
&mut self,
|
|
callback: &mut F,
|
|
old_tree: Option<&Tree>,
|
|
) -> Option<Tree> {
|
|
// A pointer to this payload is passed on every call to the `read` C function.
|
|
// The payload contains two things:
|
|
// 1. A reference to the rust `callback`.
|
|
// 2. The text that was returned from the previous call to `callback`.
|
|
// This allows the callback to return owned values like vectors.
|
|
let mut payload: (&mut F, Option<T>) = (callback, None);
|
|
|
|
// This C function is passed to Tree-sitter as the input callback.
|
|
unsafe extern "C" fn read<'a, T: AsRef<[u16]>, F: FnMut(usize, Point) -> T>(
|
|
payload: *mut c_void,
|
|
byte_offset: u32,
|
|
position: ffi::TSPoint,
|
|
bytes_read: *mut u32,
|
|
) -> *const c_char {
|
|
let (callback, text) = (payload as *mut (&mut F, Option<T>)).as_mut().unwrap();
|
|
*text = Some(callback(
|
|
(byte_offset / 2) as usize,
|
|
Point {
|
|
row: position.row as usize,
|
|
column: position.column as usize / 2,
|
|
},
|
|
));
|
|
let slice = text.as_ref().unwrap().as_ref();
|
|
*bytes_read = slice.len() as u32 * 2;
|
|
slice.as_ptr() as *const c_char
|
|
}
|
|
|
|
let c_input = ffi::TSInput {
|
|
payload: &mut payload as *mut (&mut F, Option<T>) as *mut c_void,
|
|
read: Some(read::<T, F>),
|
|
encoding: ffi::TSInputEncoding_TSInputEncodingUTF16,
|
|
};
|
|
|
|
let c_old_tree = old_tree.map_or(ptr::null_mut(), |t| t.0.as_ptr());
|
|
unsafe {
|
|
let c_new_tree = ffi::ts_parser_parse(self.0.as_ptr(), c_old_tree, c_input);
|
|
NonNull::new(c_new_tree).map(Tree)
|
|
}
|
|
}
|
|
|
|
/// Instruct the parser to start the next parse from the beginning.
|
|
///
|
|
/// If the parser previously failed because of a timeout or a cancellation, then
|
|
/// by default, it will resume where it left off on the next call to `parse` or
|
|
/// other parsing functions. If you don't want to resume, and instead intend to
|
|
/// use this parser to parse some other document, you must call `reset` first.
|
|
pub fn reset(&mut self) {
|
|
unsafe { ffi::ts_parser_reset(self.0.as_ptr()) }
|
|
}
|
|
|
|
/// Get the duration in microseconds that parsing is allowed to take.
|
|
///
|
|
/// This is set via [set_timeout_micros](Parser::set_timeout_micros).
|
|
pub fn timeout_micros(&self) -> u64 {
|
|
unsafe { ffi::ts_parser_timeout_micros(self.0.as_ptr()) }
|
|
}
|
|
|
|
/// Set the maximum duration in microseconds that parsing should be allowed to
|
|
/// take before halting.
|
|
///
|
|
/// If parsing takes longer than this, it will halt early, returning `None`.
|
|
/// See `parse` for more information.
|
|
pub fn set_timeout_micros(&mut self, timeout_micros: u64) {
|
|
unsafe { ffi::ts_parser_set_timeout_micros(self.0.as_ptr(), timeout_micros) }
|
|
}
|
|
|
|
/// Set the ranges of text that the parser should include when parsing.
|
|
///
|
|
/// By default, the parser will always include entire documents. This function
|
|
/// allows you to parse only a *portion* of a document but still return a syntax
|
|
/// tree whose ranges match up with the document as a whole. You can also pass
|
|
/// multiple disjoint ranges.
|
|
///
|
|
/// If `ranges` is empty, then the entire document will be parsed. Otherwise,
|
|
/// the given ranges must be ordered from earliest to latest in the document,
|
|
/// and they must not overlap. That is, the following must hold for all
|
|
/// `i` < `length - 1`:
|
|
/// ```text
|
|
/// ranges[i].end_byte <= ranges[i + 1].start_byte
|
|
/// ```
|
|
/// If this requirement is not satisfied, method will return IncludedRangesError
|
|
/// error with an offset in the passed ranges slice pointing to a first incorrect range.
|
|
pub fn set_included_ranges<'a>(
|
|
&mut self,
|
|
ranges: &'a [Range],
|
|
) -> Result<(), IncludedRangesError> {
|
|
let ts_ranges: Vec<ffi::TSRange> =
|
|
ranges.iter().cloned().map(|range| range.into()).collect();
|
|
let result = unsafe {
|
|
ffi::ts_parser_set_included_ranges(
|
|
self.0.as_ptr(),
|
|
ts_ranges.as_ptr(),
|
|
ts_ranges.len() as u32,
|
|
)
|
|
};
|
|
|
|
if result {
|
|
Ok(())
|
|
} else {
|
|
let mut prev_end_byte = 0;
|
|
for (i, range) in ranges.iter().enumerate() {
|
|
if range.start_byte < prev_end_byte || range.end_byte < range.start_byte {
|
|
return Err(IncludedRangesError(i));
|
|
}
|
|
prev_end_byte = range.end_byte;
|
|
}
|
|
Err(IncludedRangesError(0))
|
|
}
|
|
}
|
|
|
|
/// Get the parser's current cancellation flag pointer.
|
|
pub unsafe fn cancellation_flag(&self) -> Option<&AtomicUsize> {
|
|
(ffi::ts_parser_cancellation_flag(self.0.as_ptr()) as *const AtomicUsize).as_ref()
|
|
}
|
|
|
|
/// Set the parser's current cancellation flag pointer.
|
|
///
|
|
/// If a pointer is assigned, then the parser will periodically read from
|
|
/// this pointer during parsing. If it reads a non-zero value, it will halt early,
|
|
/// returning `None`. See [parse](Parser::parse) for more information.
|
|
pub unsafe fn set_cancellation_flag(&self, flag: Option<&AtomicUsize>) {
|
|
if let Some(flag) = flag {
|
|
ffi::ts_parser_set_cancellation_flag(
|
|
self.0.as_ptr(),
|
|
flag as *const AtomicUsize as *const usize,
|
|
);
|
|
} else {
|
|
ffi::ts_parser_set_cancellation_flag(self.0.as_ptr(), ptr::null());
|
|
}
|
|
}
|
|
}
|
|
|
|
impl Drop for Parser {
|
|
fn drop(&mut self) {
|
|
self.stop_printing_dot_graphs();
|
|
self.set_logger(None);
|
|
unsafe { ffi::ts_parser_delete(self.0.as_ptr()) }
|
|
}
|
|
}
|
|
|
|
impl Tree {
|
|
/// Get the root node of the syntax tree.
|
|
pub fn root_node(&self) -> Node {
|
|
Node::new(unsafe { ffi::ts_tree_root_node(self.0.as_ptr()) }).unwrap()
|
|
}
|
|
|
|
/// Get the language that was used to parse the syntax tree.
|
|
pub fn language(&self) -> Language {
|
|
Language(unsafe { ffi::ts_tree_language(self.0.as_ptr()) })
|
|
}
|
|
|
|
/// Edit the syntax tree to keep it in sync with source code that has been
|
|
/// edited.
|
|
///
|
|
/// You must describe the edit both in terms of byte offsets and in terms of
|
|
/// row/column coordinates.
|
|
pub fn edit(&mut self, edit: &InputEdit) {
|
|
let edit = edit.into();
|
|
unsafe { ffi::ts_tree_edit(self.0.as_ptr(), &edit) };
|
|
}
|
|
|
|
/// Create a new [TreeCursor] starting from the root of the tree.
|
|
pub fn walk(&self) -> TreeCursor {
|
|
self.root_node().walk()
|
|
}
|
|
|
|
/// Compare this old edited syntax tree to a new syntax tree representing the same
|
|
/// document, returning a sequence of ranges whose syntactic structure has changed.
|
|
///
|
|
/// For this to work correctly, this syntax tree must have been edited such that its
|
|
/// ranges match up to the new tree. Generally, you'll want to call this method right
|
|
/// after calling one of the [Parser::parse] functions. Call it on the old tree that
|
|
/// was passed to parse, and pass the new tree that was returned from `parse`.
|
|
pub fn changed_ranges(&self, other: &Tree) -> impl ExactSizeIterator<Item = Range> {
|
|
let mut count = 0;
|
|
unsafe {
|
|
let ptr = ffi::ts_tree_get_changed_ranges(
|
|
self.0.as_ptr(),
|
|
other.0.as_ptr(),
|
|
&mut count as *mut _ as *mut u32,
|
|
);
|
|
util::CBufferIter::new(ptr, count).map(|r| r.into())
|
|
}
|
|
}
|
|
}
|
|
|
|
impl fmt::Debug for Tree {
|
|
fn fmt(&self, f: &mut fmt::Formatter) -> Result<(), fmt::Error> {
|
|
write!(f, "{{Tree {:?}}}", self.root_node())
|
|
}
|
|
}
|
|
|
|
impl Drop for Tree {
|
|
fn drop(&mut self) {
|
|
unsafe { ffi::ts_tree_delete(self.0.as_ptr()) }
|
|
}
|
|
}
|
|
|
|
impl Clone for Tree {
|
|
fn clone(&self) -> Tree {
|
|
unsafe { Tree(NonNull::new_unchecked(ffi::ts_tree_copy(self.0.as_ptr()))) }
|
|
}
|
|
}
|
|
|
|
impl<'tree> Node<'tree> {
|
|
fn new(node: ffi::TSNode) -> Option<Self> {
|
|
if node.id.is_null() {
|
|
None
|
|
} else {
|
|
Some(Node(node, PhantomData))
|
|
}
|
|
}
|
|
|
|
/// Get a numeric id for this node that is unique.
|
|
///
|
|
/// Within a given syntax tree, no two nodes have the same id. However, if
|
|
/// a new tree is created based on an older tree, and a node from the old
|
|
/// tree is reused in the process, then that node will have the same id in
|
|
/// both trees.
|
|
pub fn id(&self) -> usize {
|
|
self.0.id as usize
|
|
}
|
|
|
|
/// Get this node's type as a numerical id.
|
|
pub fn kind_id(&self) -> u16 {
|
|
unsafe { ffi::ts_node_symbol(self.0) }
|
|
}
|
|
|
|
/// Get this node's type as a string.
|
|
pub fn kind(&self) -> &'static str {
|
|
unsafe { CStr::from_ptr(ffi::ts_node_type(self.0)) }
|
|
.to_str()
|
|
.unwrap()
|
|
}
|
|
|
|
/// Get the [Language] that was used to parse this node's syntax tree.
|
|
pub fn language(&self) -> Language {
|
|
Language(unsafe { ffi::ts_tree_language(self.0.tree) })
|
|
}
|
|
|
|
/// Check if this node is *named*.
|
|
///
|
|
/// Named nodes correspond to named rules in the grammar, whereas *anonymous* nodes
|
|
/// correspond to string literals in the grammar.
|
|
pub fn is_named(&self) -> bool {
|
|
unsafe { ffi::ts_node_is_named(self.0) }
|
|
}
|
|
|
|
/// Check if this node is *extra*.
|
|
///
|
|
/// Extra nodes represent things like comments, which are not required the grammar,
|
|
/// but can appear anywhere.
|
|
pub fn is_extra(&self) -> bool {
|
|
unsafe { ffi::ts_node_is_extra(self.0) }
|
|
}
|
|
|
|
/// Check if this node has been edited.
|
|
pub fn has_changes(&self) -> bool {
|
|
unsafe { ffi::ts_node_has_changes(self.0) }
|
|
}
|
|
|
|
/// Check if this node represents a syntax error or contains any syntax errors anywhere
|
|
/// within it.
|
|
pub fn has_error(&self) -> bool {
|
|
unsafe { ffi::ts_node_has_error(self.0) }
|
|
}
|
|
|
|
/// Check if this node represents a syntax error.
|
|
///
|
|
/// Syntax errors represent parts of the code that could not be incorporated into a
|
|
/// valid syntax tree.
|
|
pub fn is_error(&self) -> bool {
|
|
self.kind_id() == u16::MAX
|
|
}
|
|
|
|
/// Check if this node is *missing*.
|
|
///
|
|
/// Missing nodes are inserted by the parser in order to recover from certain kinds of
|
|
/// syntax errors.
|
|
pub fn is_missing(&self) -> bool {
|
|
unsafe { ffi::ts_node_is_missing(self.0) }
|
|
}
|
|
|
|
/// Get the byte offsets where this node starts.
|
|
pub fn start_byte(&self) -> usize {
|
|
unsafe { ffi::ts_node_start_byte(self.0) as usize }
|
|
}
|
|
|
|
/// Get the byte offsets where this node end.
|
|
pub fn end_byte(&self) -> usize {
|
|
unsafe { ffi::ts_node_end_byte(self.0) as usize }
|
|
}
|
|
|
|
/// Get the byte range of source code that this node represents.
|
|
pub fn byte_range(&self) -> std::ops::Range<usize> {
|
|
self.start_byte()..self.end_byte()
|
|
}
|
|
|
|
/// Get the range of source code that this node represents, both in terms of raw bytes
|
|
/// and of row/column coordinates.
|
|
pub fn range(&self) -> Range {
|
|
Range {
|
|
start_byte: self.start_byte(),
|
|
end_byte: self.end_byte(),
|
|
start_point: self.start_position(),
|
|
end_point: self.end_position(),
|
|
}
|
|
}
|
|
|
|
/// Get this node's start position in terms of rows and columns.
|
|
pub fn start_position(&self) -> Point {
|
|
let result = unsafe { ffi::ts_node_start_point(self.0) };
|
|
result.into()
|
|
}
|
|
|
|
/// Get this node's end position in terms of rows and columns.
|
|
pub fn end_position(&self) -> Point {
|
|
let result = unsafe { ffi::ts_node_end_point(self.0) };
|
|
result.into()
|
|
}
|
|
|
|
/// Get the node's child at the given index, where zero represents the first
|
|
/// child.
|
|
///
|
|
/// This method is fairly fast, but its cost is technically log(i), so you
|
|
/// if you might be iterating over a long list of children, you should use
|
|
/// [Node::children] instead.
|
|
pub fn child(&self, i: usize) -> Option<Self> {
|
|
Self::new(unsafe { ffi::ts_node_child(self.0, i as u32) })
|
|
}
|
|
|
|
/// Get this node's number of children.
|
|
pub fn child_count(&self) -> usize {
|
|
unsafe { ffi::ts_node_child_count(self.0) as usize }
|
|
}
|
|
|
|
/// Get this node's *named* child at the given index.
|
|
///
|
|
/// See also [Node::is_named].
|
|
/// This method is fairly fast, but its cost is technically log(i), so you
|
|
/// if you might be iterating over a long list of children, you should use
|
|
/// [Node::named_children] instead.
|
|
pub fn named_child<'a>(&'a self, i: usize) -> Option<Self> {
|
|
Self::new(unsafe { ffi::ts_node_named_child(self.0, i as u32) })
|
|
}
|
|
|
|
/// Get this node's number of *named* children.
|
|
///
|
|
/// See also [Node::is_named].
|
|
pub fn named_child_count(&self) -> usize {
|
|
unsafe { ffi::ts_node_named_child_count(self.0) as usize }
|
|
}
|
|
|
|
/// Get the first child with the given field name.
|
|
///
|
|
/// If multiple children may have the same field name, access them using
|
|
/// [children_by_field_name](Node::children_by_field_name)
|
|
pub fn child_by_field_name(&self, field_name: impl AsRef<[u8]>) -> Option<Self> {
|
|
let field_name = field_name.as_ref();
|
|
Self::new(unsafe {
|
|
ffi::ts_node_child_by_field_name(
|
|
self.0,
|
|
field_name.as_ptr() as *const c_char,
|
|
field_name.len() as u32,
|
|
)
|
|
})
|
|
}
|
|
|
|
/// Get this node's child with the given numerical field id.
|
|
///
|
|
/// See also [child_by_field_name](Node::child_by_field_name). You can convert a field name to
|
|
/// an id using [Language::field_id_for_name].
|
|
pub fn child_by_field_id(&self, field_id: u16) -> Option<Self> {
|
|
Self::new(unsafe { ffi::ts_node_child_by_field_id(self.0, field_id) })
|
|
}
|
|
|
|
/// Get the field name of this node's child at the given index.
|
|
pub fn field_name_for_child(&self, child_index: u32) -> Option<&'static str> {
|
|
unsafe {
|
|
let ptr = ffi::ts_node_field_name_for_child(self.0, child_index);
|
|
if ptr.is_null() {
|
|
None
|
|
} else {
|
|
Some(CStr::from_ptr(ptr).to_str().unwrap())
|
|
}
|
|
}
|
|
}
|
|
|
|
/// Iterate over this node's children.
|
|
///
|
|
/// A [TreeCursor] is used to retrieve the children efficiently. Obtain
|
|
/// a [TreeCursor] by calling [Tree::walk] or [Node::walk]. To avoid unnecessary
|
|
/// allocations, you should reuse the same cursor for subsequent calls to
|
|
/// this method.
|
|
///
|
|
/// If you're walking the tree recursively, you may want to use the `TreeCursor`
|
|
/// APIs directly instead.
|
|
pub fn children<'a>(
|
|
&self,
|
|
cursor: &'a mut TreeCursor<'tree>,
|
|
) -> impl ExactSizeIterator<Item = Node<'tree>> + 'a {
|
|
cursor.reset(*self);
|
|
cursor.goto_first_child();
|
|
(0..self.child_count()).into_iter().map(move |_| {
|
|
let result = cursor.node();
|
|
cursor.goto_next_sibling();
|
|
result
|
|
})
|
|
}
|
|
|
|
/// Iterate over this node's named children.
|
|
///
|
|
/// See also [Node::children].
|
|
pub fn named_children<'a>(
|
|
&self,
|
|
cursor: &'a mut TreeCursor<'tree>,
|
|
) -> impl ExactSizeIterator<Item = Node<'tree>> + 'a {
|
|
cursor.reset(*self);
|
|
cursor.goto_first_child();
|
|
(0..self.named_child_count()).into_iter().map(move |_| {
|
|
while !cursor.node().is_named() {
|
|
if !cursor.goto_next_sibling() {
|
|
break;
|
|
}
|
|
}
|
|
let result = cursor.node();
|
|
cursor.goto_next_sibling();
|
|
result
|
|
})
|
|
}
|
|
|
|
/// Iterate over this node's children with a given field name.
|
|
///
|
|
/// See also [Node::children].
|
|
pub fn children_by_field_name<'a>(
|
|
&self,
|
|
field_name: &str,
|
|
cursor: &'a mut TreeCursor<'tree>,
|
|
) -> impl Iterator<Item = Node<'tree>> + 'a {
|
|
let field_id = self.language().field_id_for_name(field_name);
|
|
self.children_by_field_id(field_id.unwrap_or(0), cursor)
|
|
}
|
|
|
|
/// Iterate over this node's children with a given field id.
|
|
///
|
|
/// See also [Node::children_by_field_name].
|
|
pub fn children_by_field_id<'a>(
|
|
&self,
|
|
field_id: u16,
|
|
cursor: &'a mut TreeCursor<'tree>,
|
|
) -> impl Iterator<Item = Node<'tree>> + 'a {
|
|
cursor.reset(*self);
|
|
cursor.goto_first_child();
|
|
let mut done = false;
|
|
iter::from_fn(move || {
|
|
while !done {
|
|
while cursor.field_id() != Some(field_id) {
|
|
if !cursor.goto_next_sibling() {
|
|
return None;
|
|
}
|
|
}
|
|
let result = cursor.node();
|
|
if !cursor.goto_next_sibling() {
|
|
done = true;
|
|
}
|
|
return Some(result);
|
|
}
|
|
None
|
|
})
|
|
}
|
|
|
|
/// Get this node's immediate parent.
|
|
pub fn parent(&self) -> Option<Self> {
|
|
Self::new(unsafe { ffi::ts_node_parent(self.0) })
|
|
}
|
|
|
|
/// Get this node's next sibling.
|
|
pub fn next_sibling(&self) -> Option<Self> {
|
|
Self::new(unsafe { ffi::ts_node_next_sibling(self.0) })
|
|
}
|
|
|
|
/// Get this node's previous sibling.
|
|
pub fn prev_sibling(&self) -> Option<Self> {
|
|
Self::new(unsafe { ffi::ts_node_prev_sibling(self.0) })
|
|
}
|
|
|
|
/// Get this node's next named sibling.
|
|
pub fn next_named_sibling(&self) -> Option<Self> {
|
|
Self::new(unsafe { ffi::ts_node_next_named_sibling(self.0) })
|
|
}
|
|
|
|
/// Get this node's previous named sibling.
|
|
pub fn prev_named_sibling(&self) -> Option<Self> {
|
|
Self::new(unsafe { ffi::ts_node_prev_named_sibling(self.0) })
|
|
}
|
|
|
|
/// Get the smallest node within this node that spans the given range.
|
|
pub fn descendant_for_byte_range(&self, start: usize, end: usize) -> Option<Self> {
|
|
Self::new(unsafe {
|
|
ffi::ts_node_descendant_for_byte_range(self.0, start as u32, end as u32)
|
|
})
|
|
}
|
|
|
|
/// Get the smallest named node within this node that spans the given range.
|
|
pub fn named_descendant_for_byte_range(&self, start: usize, end: usize) -> Option<Self> {
|
|
Self::new(unsafe {
|
|
ffi::ts_node_named_descendant_for_byte_range(self.0, start as u32, end as u32)
|
|
})
|
|
}
|
|
|
|
/// Get the smallest node within this node that spans the given range.
|
|
pub fn descendant_for_point_range(&self, start: Point, end: Point) -> Option<Self> {
|
|
Self::new(unsafe {
|
|
ffi::ts_node_descendant_for_point_range(self.0, start.into(), end.into())
|
|
})
|
|
}
|
|
|
|
/// Get the smallest named node within this node that spans the given range.
|
|
pub fn named_descendant_for_point_range(&self, start: Point, end: Point) -> Option<Self> {
|
|
Self::new(unsafe {
|
|
ffi::ts_node_named_descendant_for_point_range(self.0, start.into(), end.into())
|
|
})
|
|
}
|
|
|
|
pub fn to_sexp(&self) -> String {
|
|
let c_string = unsafe { ffi::ts_node_string(self.0) };
|
|
let result = unsafe { CStr::from_ptr(c_string) }
|
|
.to_str()
|
|
.unwrap()
|
|
.to_string();
|
|
unsafe { util::free_ptr(c_string as *mut c_void) };
|
|
result
|
|
}
|
|
|
|
pub fn utf8_text<'a>(&self, source: &'a [u8]) -> Result<&'a str, str::Utf8Error> {
|
|
str::from_utf8(&source[self.start_byte()..self.end_byte()])
|
|
}
|
|
|
|
pub fn utf16_text<'a>(&self, source: &'a [u16]) -> &'a [u16] {
|
|
&source.as_ref()[self.start_byte()..self.end_byte()]
|
|
}
|
|
|
|
/// Create a new [TreeCursor] starting from this node.
|
|
pub fn walk(&self) -> TreeCursor<'tree> {
|
|
TreeCursor(unsafe { ffi::ts_tree_cursor_new(self.0) }, PhantomData)
|
|
}
|
|
|
|
/// Edit this node to keep it in-sync with source code that has been edited.
|
|
///
|
|
/// This function is only rarely needed. When you edit a syntax tree with the
|
|
/// [Tree::edit] method, all of the nodes that you retrieve from the tree
|
|
/// afterward will already reflect the edit. You only need to use [Node::edit]
|
|
/// when you have a specific [Node] instance that you want to keep and continue
|
|
/// to use after an edit.
|
|
pub fn edit(&mut self, edit: &InputEdit) {
|
|
let edit = edit.into();
|
|
unsafe { ffi::ts_node_edit(&mut self.0 as *mut ffi::TSNode, &edit) }
|
|
}
|
|
}
|
|
|
|
impl<'a> PartialEq for Node<'a> {
|
|
fn eq(&self, other: &Self) -> bool {
|
|
self.0.id == other.0.id
|
|
}
|
|
}
|
|
|
|
impl<'a> Eq for Node<'a> {}
|
|
|
|
impl<'a> hash::Hash for Node<'a> {
|
|
fn hash<H: hash::Hasher>(&self, state: &mut H) {
|
|
self.0.id.hash(state);
|
|
self.0.context[0].hash(state);
|
|
self.0.context[1].hash(state);
|
|
self.0.context[2].hash(state);
|
|
self.0.context[3].hash(state);
|
|
}
|
|
}
|
|
|
|
impl<'a> fmt::Debug for Node<'a> {
|
|
fn fmt(&self, f: &mut fmt::Formatter) -> Result<(), fmt::Error> {
|
|
write!(
|
|
f,
|
|
"{{Node {} {} - {}}}",
|
|
self.kind(),
|
|
self.start_position(),
|
|
self.end_position()
|
|
)
|
|
}
|
|
}
|
|
|
|
impl<'a> TreeCursor<'a> {
|
|
/// Get the tree cursor's current [Node].
|
|
pub fn node(&self) -> Node<'a> {
|
|
Node(
|
|
unsafe { ffi::ts_tree_cursor_current_node(&self.0) },
|
|
PhantomData,
|
|
)
|
|
}
|
|
|
|
/// Get the numerical field id of this tree cursor's current node.
|
|
///
|
|
/// See also [field_name](TreeCursor::field_name).
|
|
pub fn field_id(&self) -> Option<u16> {
|
|
unsafe {
|
|
let id = ffi::ts_tree_cursor_current_field_id(&self.0);
|
|
if id == 0 {
|
|
None
|
|
} else {
|
|
Some(id)
|
|
}
|
|
}
|
|
}
|
|
|
|
/// Get the field name of this tree cursor's current node.
|
|
pub fn field_name(&self) -> Option<&'static str> {
|
|
unsafe {
|
|
let ptr = ffi::ts_tree_cursor_current_field_name(&self.0);
|
|
if ptr.is_null() {
|
|
None
|
|
} else {
|
|
Some(CStr::from_ptr(ptr).to_str().unwrap())
|
|
}
|
|
}
|
|
}
|
|
|
|
/// Move this cursor to the first child of its current node.
|
|
///
|
|
/// This returns `true` if the cursor successfully moved, and returns `false`
|
|
/// if there were no children.
|
|
pub fn goto_first_child(&mut self) -> bool {
|
|
return unsafe { ffi::ts_tree_cursor_goto_first_child(&mut self.0) };
|
|
}
|
|
|
|
/// Move this cursor to the parent of its current node.
|
|
///
|
|
/// This returns `true` if the cursor successfully moved, and returns `false`
|
|
/// if there was no parent node (the cursor was already on the root node).
|
|
pub fn goto_parent(&mut self) -> bool {
|
|
return unsafe { ffi::ts_tree_cursor_goto_parent(&mut self.0) };
|
|
}
|
|
|
|
/// Move this cursor to the next sibling of its current node.
|
|
///
|
|
/// This returns `true` if the cursor successfully moved, and returns `false`
|
|
/// if there was no next sibling node.
|
|
pub fn goto_next_sibling(&mut self) -> bool {
|
|
return unsafe { ffi::ts_tree_cursor_goto_next_sibling(&mut self.0) };
|
|
}
|
|
|
|
/// Move this cursor to the first child of its current node that extends beyond
|
|
/// the given byte offset.
|
|
///
|
|
/// This returns the index of the child node if one was found, and returns `None`
|
|
/// if no such child was found.
|
|
pub fn goto_first_child_for_byte(&mut self, index: usize) -> Option<usize> {
|
|
let result =
|
|
unsafe { ffi::ts_tree_cursor_goto_first_child_for_byte(&mut self.0, index as u32) };
|
|
if result < 0 {
|
|
None
|
|
} else {
|
|
Some(result as usize)
|
|
}
|
|
}
|
|
|
|
/// Move this cursor to the first child of its current node that extends beyond
|
|
/// the given byte offset.
|
|
///
|
|
/// This returns the index of the child node if one was found, and returns `None`
|
|
/// if no such child was found.
|
|
pub fn goto_first_child_for_point(&mut self, point: Point) -> Option<usize> {
|
|
let result =
|
|
unsafe { ffi::ts_tree_cursor_goto_first_child_for_point(&mut self.0, point.into()) };
|
|
if result < 0 {
|
|
None
|
|
} else {
|
|
Some(result as usize)
|
|
}
|
|
}
|
|
|
|
/// Re-initialize this tree cursor to start at a different node.
|
|
pub fn reset(&mut self, node: Node<'a>) {
|
|
unsafe { ffi::ts_tree_cursor_reset(&mut self.0, node.0) };
|
|
}
|
|
}
|
|
|
|
impl<'a> Clone for TreeCursor<'a> {
|
|
fn clone(&self) -> Self {
|
|
TreeCursor(unsafe { ffi::ts_tree_cursor_copy(&self.0) }, PhantomData)
|
|
}
|
|
}
|
|
|
|
impl<'a> Drop for TreeCursor<'a> {
|
|
fn drop(&mut self) {
|
|
unsafe { ffi::ts_tree_cursor_delete(&mut self.0) }
|
|
}
|
|
}
|
|
|
|
impl Query {
|
|
/// Create a new query from a string containing one or more S-expression
|
|
/// patterns.
|
|
///
|
|
/// The query is associated with a particular language, and can only be run
|
|
/// on syntax nodes parsed with that language. References to Queries can be
|
|
/// shared between multiple threads.
|
|
pub fn new(language: Language, source: &str) -> Result<Self, QueryError> {
|
|
let mut error_offset = 0u32;
|
|
let mut error_type: ffi::TSQueryError = 0;
|
|
let bytes = source.as_bytes();
|
|
|
|
// Compile the query.
|
|
let ptr = unsafe {
|
|
ffi::ts_query_new(
|
|
language.0,
|
|
bytes.as_ptr() as *const c_char,
|
|
bytes.len() as u32,
|
|
&mut error_offset as *mut u32,
|
|
&mut error_type as *mut ffi::TSQueryError,
|
|
)
|
|
};
|
|
|
|
// On failure, build an error based on the error code and offset.
|
|
if ptr.is_null() {
|
|
let offset = error_offset as usize;
|
|
let mut line_start = 0;
|
|
let mut row = 0;
|
|
let mut line_containing_error = None;
|
|
for line in source.split("\n") {
|
|
let line_end = line_start + line.len() + 1;
|
|
if line_end > offset {
|
|
line_containing_error = Some(line);
|
|
break;
|
|
}
|
|
line_start = line_end;
|
|
row += 1;
|
|
}
|
|
let column = offset - line_start;
|
|
|
|
let kind;
|
|
let message;
|
|
match error_type {
|
|
// Error types that report names
|
|
ffi::TSQueryError_TSQueryErrorNodeType
|
|
| ffi::TSQueryError_TSQueryErrorField
|
|
| ffi::TSQueryError_TSQueryErrorCapture => {
|
|
let suffix = source.split_at(offset).1;
|
|
let end_offset = suffix
|
|
.find(|c| !char::is_alphanumeric(c) && c != '_' && c != '-')
|
|
.unwrap_or(source.len());
|
|
message = suffix.split_at(end_offset).0.to_string();
|
|
kind = match error_type {
|
|
ffi::TSQueryError_TSQueryErrorNodeType => QueryErrorKind::NodeType,
|
|
ffi::TSQueryError_TSQueryErrorField => QueryErrorKind::Field,
|
|
ffi::TSQueryError_TSQueryErrorCapture => QueryErrorKind::Capture,
|
|
_ => unreachable!(),
|
|
};
|
|
}
|
|
|
|
// Error types that report positions
|
|
_ => {
|
|
message = if let Some(line) = line_containing_error {
|
|
line.to_string() + "\n" + &" ".repeat(offset - line_start) + "^"
|
|
} else {
|
|
"Unexpected EOF".to_string()
|
|
};
|
|
kind = match error_type {
|
|
ffi::TSQueryError_TSQueryErrorStructure => QueryErrorKind::Structure,
|
|
_ => QueryErrorKind::Syntax,
|
|
};
|
|
}
|
|
};
|
|
|
|
return Err(QueryError {
|
|
row,
|
|
column,
|
|
offset,
|
|
kind,
|
|
message,
|
|
});
|
|
}
|
|
|
|
let string_count = unsafe { ffi::ts_query_string_count(ptr) };
|
|
let capture_count = unsafe { ffi::ts_query_capture_count(ptr) };
|
|
let pattern_count = unsafe { ffi::ts_query_pattern_count(ptr) as usize };
|
|
let mut result = Query {
|
|
ptr: unsafe { NonNull::new_unchecked(ptr) },
|
|
capture_names: Vec::with_capacity(capture_count as usize),
|
|
text_predicates: Vec::with_capacity(pattern_count),
|
|
property_predicates: Vec::with_capacity(pattern_count),
|
|
property_settings: Vec::with_capacity(pattern_count),
|
|
general_predicates: Vec::with_capacity(pattern_count),
|
|
};
|
|
|
|
// Build a vector of strings to store the capture names.
|
|
for i in 0..capture_count {
|
|
unsafe {
|
|
let mut length = 0u32;
|
|
let name =
|
|
ffi::ts_query_capture_name_for_id(ptr, i, &mut length as *mut u32) as *const u8;
|
|
let name = slice::from_raw_parts(name, length as usize);
|
|
let name = str::from_utf8_unchecked(name);
|
|
result.capture_names.push(name.to_string());
|
|
}
|
|
}
|
|
|
|
// Build a vector of strings to represent literal values used in predicates.
|
|
let string_values = (0..string_count)
|
|
.map(|i| unsafe {
|
|
let mut length = 0u32;
|
|
let value =
|
|
ffi::ts_query_string_value_for_id(ptr, i as u32, &mut length as *mut u32)
|
|
as *const u8;
|
|
let value = slice::from_raw_parts(value, length as usize);
|
|
let value = str::from_utf8_unchecked(value);
|
|
value.to_string()
|
|
})
|
|
.collect::<Vec<_>>();
|
|
|
|
// Build a vector of predicates for each pattern.
|
|
for i in 0..pattern_count {
|
|
let predicate_steps = unsafe {
|
|
let mut length = 0u32;
|
|
let raw_predicates =
|
|
ffi::ts_query_predicates_for_pattern(ptr, i as u32, &mut length as *mut u32);
|
|
if length > 0 {
|
|
slice::from_raw_parts(raw_predicates, length as usize)
|
|
} else {
|
|
&[]
|
|
}
|
|
};
|
|
|
|
let byte_offset = unsafe { ffi::ts_query_start_byte_for_pattern(ptr, i as u32) };
|
|
let row = source
|
|
.char_indices()
|
|
.take_while(|(i, _)| *i < byte_offset as usize)
|
|
.filter(|(_, c)| *c == '\n')
|
|
.count();
|
|
|
|
let type_done = ffi::TSQueryPredicateStepType_TSQueryPredicateStepTypeDone;
|
|
let type_capture = ffi::TSQueryPredicateStepType_TSQueryPredicateStepTypeCapture;
|
|
let type_string = ffi::TSQueryPredicateStepType_TSQueryPredicateStepTypeString;
|
|
|
|
let mut text_predicates = Vec::new();
|
|
let mut property_predicates = Vec::new();
|
|
let mut property_settings = Vec::new();
|
|
let mut general_predicates = Vec::new();
|
|
for p in predicate_steps.split(|s| s.type_ == type_done) {
|
|
if p.is_empty() {
|
|
continue;
|
|
}
|
|
|
|
if p[0].type_ != type_string {
|
|
return Err(predicate_error(
|
|
row,
|
|
format!(
|
|
"Expected predicate to start with a function name. Got @{}.",
|
|
result.capture_names[p[0].value_id as usize],
|
|
),
|
|
));
|
|
}
|
|
|
|
// Build a predicate for each of the known predicate function names.
|
|
let operator_name = &string_values[p[0].value_id as usize];
|
|
match operator_name.as_str() {
|
|
"eq?" | "not-eq?" => {
|
|
if p.len() != 3 {
|
|
return Err(predicate_error(
|
|
row,
|
|
format!(
|
|
"Wrong number of arguments to #eq? predicate. Expected 2, got {}.",
|
|
p.len() - 1
|
|
),
|
|
));
|
|
}
|
|
if p[1].type_ != type_capture {
|
|
return Err(predicate_error(row, format!(
|
|
"First argument to #eq? predicate must be a capture name. Got literal \"{}\".",
|
|
string_values[p[1].value_id as usize],
|
|
)));
|
|
}
|
|
|
|
let is_positive = operator_name == "eq?";
|
|
text_predicates.push(if p[2].type_ == type_capture {
|
|
TextPredicate::CaptureEqCapture(
|
|
p[1].value_id,
|
|
p[2].value_id,
|
|
is_positive,
|
|
)
|
|
} else {
|
|
TextPredicate::CaptureEqString(
|
|
p[1].value_id,
|
|
string_values[p[2].value_id as usize].clone(),
|
|
is_positive,
|
|
)
|
|
});
|
|
}
|
|
|
|
"match?" | "not-match?" => {
|
|
if p.len() != 3 {
|
|
return Err(predicate_error(row, format!(
|
|
"Wrong number of arguments to #match? predicate. Expected 2, got {}.",
|
|
p.len() - 1
|
|
)));
|
|
}
|
|
if p[1].type_ != type_capture {
|
|
return Err(predicate_error(row, format!(
|
|
"First argument to #match? predicate must be a capture name. Got literal \"{}\".",
|
|
string_values[p[1].value_id as usize],
|
|
)));
|
|
}
|
|
if p[2].type_ == type_capture {
|
|
return Err(predicate_error(row, format!(
|
|
"Second argument to #match? predicate must be a literal. Got capture @{}.",
|
|
result.capture_names[p[2].value_id as usize],
|
|
)));
|
|
}
|
|
|
|
let is_positive = operator_name == "match?";
|
|
let regex = &string_values[p[2].value_id as usize];
|
|
text_predicates.push(TextPredicate::CaptureMatchString(
|
|
p[1].value_id,
|
|
regex::bytes::Regex::new(regex).map_err(|_| {
|
|
predicate_error(row, format!("Invalid regex '{}'", regex))
|
|
})?,
|
|
is_positive,
|
|
));
|
|
}
|
|
|
|
"set!" => property_settings.push(Self::parse_property(
|
|
row,
|
|
&operator_name,
|
|
&result.capture_names,
|
|
&string_values,
|
|
&p[1..],
|
|
)?),
|
|
|
|
"is?" | "is-not?" => property_predicates.push((
|
|
Self::parse_property(
|
|
row,
|
|
&operator_name,
|
|
&result.capture_names,
|
|
&string_values,
|
|
&p[1..],
|
|
)?,
|
|
operator_name == "is?",
|
|
)),
|
|
|
|
_ => general_predicates.push(QueryPredicate {
|
|
operator: operator_name.clone().into_boxed_str(),
|
|
args: p[1..]
|
|
.iter()
|
|
.map(|a| {
|
|
if a.type_ == type_capture {
|
|
QueryPredicateArg::Capture(a.value_id)
|
|
} else {
|
|
QueryPredicateArg::String(
|
|
string_values[a.value_id as usize].clone().into_boxed_str(),
|
|
)
|
|
}
|
|
})
|
|
.collect(),
|
|
}),
|
|
}
|
|
}
|
|
|
|
result
|
|
.text_predicates
|
|
.push(text_predicates.into_boxed_slice());
|
|
result
|
|
.property_predicates
|
|
.push(property_predicates.into_boxed_slice());
|
|
result
|
|
.property_settings
|
|
.push(property_settings.into_boxed_slice());
|
|
result
|
|
.general_predicates
|
|
.push(general_predicates.into_boxed_slice());
|
|
}
|
|
Ok(result)
|
|
}
|
|
|
|
/// Get the byte offset where the given pattern starts in the query's source.
|
|
pub fn start_byte_for_pattern(&self, pattern_index: usize) -> usize {
|
|
if pattern_index >= self.text_predicates.len() {
|
|
panic!(
|
|
"Pattern index is {} but the pattern count is {}",
|
|
pattern_index,
|
|
self.text_predicates.len(),
|
|
);
|
|
}
|
|
unsafe {
|
|
ffi::ts_query_start_byte_for_pattern(self.ptr.as_ptr(), pattern_index as u32) as usize
|
|
}
|
|
}
|
|
|
|
/// Get the number of patterns in the query.
|
|
pub fn pattern_count(&self) -> usize {
|
|
unsafe { ffi::ts_query_pattern_count(self.ptr.as_ptr()) as usize }
|
|
}
|
|
|
|
/// Get the names of the captures used in the query.
|
|
pub fn capture_names(&self) -> &[String] {
|
|
&self.capture_names
|
|
}
|
|
|
|
/// Get the properties that are checked for the given pattern index.
|
|
///
|
|
/// This includes predicates with the operators `is?` and `is-not?`.
|
|
pub fn property_predicates(&self, index: usize) -> &[(QueryProperty, bool)] {
|
|
&self.property_predicates[index]
|
|
}
|
|
|
|
/// Get the properties that are set for the given pattern index.
|
|
///
|
|
/// This includes predicates with the operator `set!`.
|
|
pub fn property_settings(&self, index: usize) -> &[QueryProperty] {
|
|
&self.property_settings[index]
|
|
}
|
|
|
|
/// Get the other user-defined predicates associated with the given index.
|
|
///
|
|
/// This includes predicate with operators other than:
|
|
/// * `match?`
|
|
/// * `eq?` and `not-eq?`
|
|
/// * `is?` and `is-not?`
|
|
/// * `set!`
|
|
pub fn general_predicates(&self, index: usize) -> &[QueryPredicate] {
|
|
&self.general_predicates[index]
|
|
}
|
|
|
|
/// Disable a certain capture within a query.
|
|
///
|
|
/// This prevents the capture from being returned in matches, and also avoids any
|
|
/// resource usage associated with recording the capture.
|
|
pub fn disable_capture(&mut self, name: &str) {
|
|
unsafe {
|
|
ffi::ts_query_disable_capture(
|
|
self.ptr.as_ptr(),
|
|
name.as_bytes().as_ptr() as *const c_char,
|
|
name.len() as u32,
|
|
);
|
|
}
|
|
}
|
|
|
|
/// Disable a certain pattern within a query.
|
|
///
|
|
/// This prevents the pattern from matching, and also avoids any resource usage
|
|
/// associated with the pattern.
|
|
pub fn disable_pattern(&mut self, index: usize) {
|
|
unsafe { ffi::ts_query_disable_pattern(self.ptr.as_ptr(), index as u32) }
|
|
}
|
|
|
|
/// Check if a given step in a query is 'definite'.
|
|
///
|
|
/// A query step is 'definite' if its parent pattern will be guaranteed to match
|
|
/// successfully once it reaches the step.
|
|
pub fn step_is_definite(&self, byte_offset: usize) -> bool {
|
|
unsafe { ffi::ts_query_step_is_definite(self.ptr.as_ptr(), byte_offset as u32) }
|
|
}
|
|
|
|
fn parse_property(
|
|
row: usize,
|
|
function_name: &str,
|
|
capture_names: &[String],
|
|
string_values: &[String],
|
|
args: &[ffi::TSQueryPredicateStep],
|
|
) -> Result<QueryProperty, QueryError> {
|
|
if args.len() == 0 || args.len() > 3 {
|
|
return Err(predicate_error(
|
|
row,
|
|
format!(
|
|
"Wrong number of arguments to {} predicate. Expected 1 to 3, got {}.",
|
|
function_name,
|
|
args.len(),
|
|
),
|
|
));
|
|
}
|
|
|
|
let mut capture_id = None;
|
|
let mut key = None;
|
|
let mut value = None;
|
|
|
|
for arg in args {
|
|
if arg.type_ == ffi::TSQueryPredicateStepType_TSQueryPredicateStepTypeCapture {
|
|
if capture_id.is_some() {
|
|
return Err(predicate_error(
|
|
row,
|
|
format!(
|
|
"Invalid arguments to {} predicate. Unexpected second capture name @{}",
|
|
function_name, capture_names[arg.value_id as usize]
|
|
),
|
|
));
|
|
}
|
|
capture_id = Some(arg.value_id as usize);
|
|
} else if key.is_none() {
|
|
key = Some(&string_values[arg.value_id as usize]);
|
|
} else if value.is_none() {
|
|
value = Some(string_values[arg.value_id as usize].as_str());
|
|
} else {
|
|
return Err(predicate_error(
|
|
row,
|
|
format!(
|
|
"Invalid arguments to {} predicate. Unexpected third argument @{}",
|
|
function_name, string_values[arg.value_id as usize]
|
|
),
|
|
));
|
|
}
|
|
}
|
|
|
|
if let Some(key) = key {
|
|
Ok(QueryProperty::new(key, value, capture_id))
|
|
} else {
|
|
return Err(predicate_error(
|
|
row,
|
|
format!(
|
|
"Invalid arguments to {} predicate. Missing key argument",
|
|
function_name,
|
|
),
|
|
));
|
|
}
|
|
}
|
|
}
|
|
|
|
impl QueryCursor {
|
|
/// Create a new cursor for executing a given query.
|
|
///
|
|
/// The cursor stores the state that is needed to iteratively search for matches.
|
|
pub fn new() -> Self {
|
|
QueryCursor {
|
|
ptr: unsafe { NonNull::new_unchecked(ffi::ts_query_cursor_new()) },
|
|
}
|
|
}
|
|
|
|
/// Check if, on its last execution, this cursor exceeded its maximum number of
|
|
/// in-progress matches.
|
|
pub fn did_exceed_match_limit(&self) -> bool {
|
|
unsafe { ffi::ts_query_cursor_did_exceed_match_limit(self.ptr.as_ptr()) }
|
|
}
|
|
|
|
/// Iterate over all of the matches in the order that they were found.
|
|
///
|
|
/// Each match contains the index of the pattern that matched, and a list of captures.
|
|
/// Because multiple patterns can match the same set of nodes, one match may contain
|
|
/// captures that appear *before* some of the captures from a previous match.
|
|
pub fn matches<'a, 'tree: 'a, T: TextProvider<'a> + 'a>(
|
|
&'a mut self,
|
|
query: &'a Query,
|
|
node: Node<'tree>,
|
|
text_provider: T,
|
|
) -> QueryMatches<'a, 'tree, T> {
|
|
let ptr = self.ptr.as_ptr();
|
|
unsafe { ffi::ts_query_cursor_exec(ptr, query.ptr.as_ptr(), node.0) };
|
|
QueryMatches {
|
|
ptr,
|
|
query,
|
|
text_provider,
|
|
buffer1: Default::default(),
|
|
buffer2: Default::default(),
|
|
_tree: PhantomData,
|
|
}
|
|
}
|
|
|
|
/// Iterate over all of the individual captures in the order that they appear.
|
|
///
|
|
/// This is useful if don't care about which pattern matched, and just want a single,
|
|
/// ordered sequence of captures.
|
|
pub fn captures<'a, 'tree: 'a, T: TextProvider<'a> + 'a>(
|
|
&'a mut self,
|
|
query: &'a Query,
|
|
node: Node<'tree>,
|
|
text_provider: T,
|
|
) -> QueryCaptures<'a, 'tree, T> {
|
|
let ptr = self.ptr.as_ptr();
|
|
unsafe { ffi::ts_query_cursor_exec(self.ptr.as_ptr(), query.ptr.as_ptr(), node.0) };
|
|
QueryCaptures {
|
|
ptr,
|
|
query,
|
|
text_provider,
|
|
buffer1: Default::default(),
|
|
buffer2: Default::default(),
|
|
_tree: PhantomData,
|
|
}
|
|
}
|
|
|
|
/// Set the range in which the query will be executed, in terms of byte offsets.
|
|
pub fn set_byte_range(&mut self, start: usize, end: usize) -> &mut Self {
|
|
unsafe {
|
|
ffi::ts_query_cursor_set_byte_range(self.ptr.as_ptr(), start as u32, end as u32);
|
|
}
|
|
self
|
|
}
|
|
|
|
/// Set the range in which the query will be executed, in terms of rows and columns.
|
|
pub fn set_point_range(&mut self, start: Point, end: Point) -> &mut Self {
|
|
unsafe {
|
|
ffi::ts_query_cursor_set_point_range(self.ptr.as_ptr(), start.into(), end.into());
|
|
}
|
|
self
|
|
}
|
|
}
|
|
|
|
impl<'a, 'tree> QueryMatch<'a, 'tree> {
|
|
pub fn remove(self) {
|
|
unsafe { ffi::ts_query_cursor_remove_match(self.cursor, self.id) }
|
|
}
|
|
|
|
fn new(m: ffi::TSQueryMatch, cursor: *mut ffi::TSQueryCursor) -> Self {
|
|
QueryMatch {
|
|
cursor,
|
|
id: m.id,
|
|
pattern_index: m.pattern_index as usize,
|
|
captures: if m.capture_count > 0 {
|
|
unsafe {
|
|
slice::from_raw_parts(
|
|
m.captures as *const QueryCapture<'tree>,
|
|
m.capture_count as usize,
|
|
)
|
|
}
|
|
} else {
|
|
&[]
|
|
},
|
|
}
|
|
}
|
|
|
|
fn satisfies_text_predicates(
|
|
&self,
|
|
query: &Query,
|
|
buffer1: &mut Vec<u8>,
|
|
buffer2: &mut Vec<u8>,
|
|
text_provider: &mut impl TextProvider<'a>,
|
|
) -> bool {
|
|
fn get_text<'a, 'b: 'a, I: Iterator<Item = &'b [u8]>>(
|
|
buffer: &'a mut Vec<u8>,
|
|
mut chunks: I,
|
|
) -> &'a [u8] {
|
|
let first_chunk = chunks.next().unwrap_or(&[]);
|
|
if let Some(next_chunk) = chunks.next() {
|
|
buffer.clear();
|
|
buffer.extend_from_slice(first_chunk);
|
|
buffer.extend_from_slice(next_chunk);
|
|
for chunk in chunks {
|
|
buffer.extend_from_slice(chunk);
|
|
}
|
|
buffer.as_slice()
|
|
} else {
|
|
first_chunk
|
|
}
|
|
}
|
|
|
|
query.text_predicates[self.pattern_index]
|
|
.iter()
|
|
.all(|predicate| match predicate {
|
|
TextPredicate::CaptureEqCapture(i, j, is_positive) => {
|
|
let node1 = self.capture_for_index(*i).unwrap();
|
|
let node2 = self.capture_for_index(*j).unwrap();
|
|
let text1 = get_text(buffer1, text_provider.text(node1));
|
|
let text2 = get_text(buffer2, text_provider.text(node2));
|
|
(text1 == text2) == *is_positive
|
|
}
|
|
TextPredicate::CaptureEqString(i, s, is_positive) => {
|
|
let node = self.capture_for_index(*i).unwrap();
|
|
let text = get_text(buffer1, text_provider.text(node));
|
|
(text == s.as_bytes()) == *is_positive
|
|
}
|
|
TextPredicate::CaptureMatchString(i, r, is_positive) => {
|
|
let node = self.capture_for_index(*i).unwrap();
|
|
let text = get_text(buffer1, text_provider.text(node));
|
|
r.is_match(text) == *is_positive
|
|
}
|
|
})
|
|
}
|
|
|
|
fn capture_for_index(&self, capture_index: u32) -> Option<Node<'tree>> {
|
|
for c in self.captures {
|
|
if c.index == capture_index {
|
|
return Some(c.node);
|
|
}
|
|
}
|
|
None
|
|
}
|
|
}
|
|
|
|
impl QueryProperty {
|
|
pub fn new(key: &str, value: Option<&str>, capture_id: Option<usize>) -> Self {
|
|
QueryProperty {
|
|
capture_id,
|
|
key: key.to_string().into_boxed_str(),
|
|
value: value.map(|s| s.to_string().into_boxed_str()),
|
|
}
|
|
}
|
|
}
|
|
|
|
impl<'a, 'tree, T: TextProvider<'a>> Iterator for QueryMatches<'a, 'tree, T> {
|
|
type Item = QueryMatch<'a, 'tree>;
|
|
|
|
fn next(&mut self) -> Option<Self::Item> {
|
|
unsafe {
|
|
loop {
|
|
let mut m = MaybeUninit::<ffi::TSQueryMatch>::uninit();
|
|
if ffi::ts_query_cursor_next_match(self.ptr, m.as_mut_ptr()) {
|
|
let result = QueryMatch::new(m.assume_init(), self.ptr);
|
|
if result.satisfies_text_predicates(
|
|
self.query,
|
|
&mut self.buffer1,
|
|
&mut self.buffer2,
|
|
&mut self.text_provider,
|
|
) {
|
|
return Some(result);
|
|
}
|
|
} else {
|
|
return None;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
impl<'a, 'tree, T: TextProvider<'a>> QueryCaptures<'a, 'tree, T> {
|
|
pub fn set_byte_range(&mut self, start: usize, end: usize) {
|
|
unsafe {
|
|
ffi::ts_query_cursor_set_byte_range(self.ptr, start as u32, end as u32);
|
|
}
|
|
}
|
|
}
|
|
|
|
impl<'a, 'tree, T: TextProvider<'a>> Iterator for QueryCaptures<'a, 'tree, T> {
|
|
type Item = (QueryMatch<'a, 'tree>, usize);
|
|
|
|
fn next(&mut self) -> Option<Self::Item> {
|
|
unsafe {
|
|
loop {
|
|
let mut capture_index = 0u32;
|
|
let mut m = MaybeUninit::<ffi::TSQueryMatch>::uninit();
|
|
if ffi::ts_query_cursor_next_capture(
|
|
self.ptr,
|
|
m.as_mut_ptr(),
|
|
&mut capture_index as *mut u32,
|
|
) {
|
|
let result = QueryMatch::new(m.assume_init(), self.ptr);
|
|
if result.satisfies_text_predicates(
|
|
self.query,
|
|
&mut self.buffer1,
|
|
&mut self.buffer2,
|
|
&mut self.text_provider,
|
|
) {
|
|
return Some((result, capture_index as usize));
|
|
} else {
|
|
result.remove();
|
|
}
|
|
} else {
|
|
return None;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
impl<'cursor, 'tree> fmt::Debug for QueryMatch<'cursor, 'tree> {
|
|
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
|
|
write!(
|
|
f,
|
|
"QueryMatch {{ id: {}, pattern_index: {}, captures: {:?} }}",
|
|
self.id, self.pattern_index, self.captures
|
|
)
|
|
}
|
|
}
|
|
|
|
impl<'a, F, I> TextProvider<'a> for F
|
|
where
|
|
F: FnMut(Node) -> I,
|
|
I: Iterator<Item = &'a [u8]> + 'a,
|
|
{
|
|
type I = I;
|
|
|
|
fn text(&mut self, node: Node) -> Self::I {
|
|
(self)(node)
|
|
}
|
|
}
|
|
|
|
impl<'a> TextProvider<'a> for &'a [u8] {
|
|
type I = std::option::IntoIter<&'a [u8]>;
|
|
|
|
fn text(&mut self, node: Node) -> Self::I {
|
|
Some(&self[node.byte_range()]).into_iter()
|
|
}
|
|
}
|
|
|
|
impl PartialEq for Query {
|
|
fn eq(&self, other: &Self) -> bool {
|
|
self.ptr == other.ptr
|
|
}
|
|
}
|
|
|
|
impl Drop for Query {
|
|
fn drop(&mut self) {
|
|
unsafe { ffi::ts_query_delete(self.ptr.as_ptr()) }
|
|
}
|
|
}
|
|
|
|
impl Drop for QueryCursor {
|
|
fn drop(&mut self) {
|
|
unsafe { ffi::ts_query_cursor_delete(self.ptr.as_ptr()) }
|
|
}
|
|
}
|
|
|
|
impl Point {
|
|
pub fn new(row: usize, column: usize) -> Self {
|
|
Point { row, column }
|
|
}
|
|
}
|
|
|
|
impl fmt::Display for Point {
|
|
fn fmt(&self, f: &mut fmt::Formatter) -> Result<(), fmt::Error> {
|
|
write!(f, "({}, {})", self.row, self.column)
|
|
}
|
|
}
|
|
|
|
impl Into<ffi::TSPoint> for Point {
|
|
fn into(self) -> ffi::TSPoint {
|
|
ffi::TSPoint {
|
|
row: self.row as u32,
|
|
column: self.column as u32,
|
|
}
|
|
}
|
|
}
|
|
|
|
impl From<ffi::TSPoint> for Point {
|
|
fn from(point: ffi::TSPoint) -> Self {
|
|
Self {
|
|
row: point.row as usize,
|
|
column: point.column as usize,
|
|
}
|
|
}
|
|
}
|
|
|
|
impl Into<ffi::TSRange> for Range {
|
|
fn into(self) -> ffi::TSRange {
|
|
ffi::TSRange {
|
|
start_byte: self.start_byte as u32,
|
|
end_byte: self.end_byte as u32,
|
|
start_point: self.start_point.into(),
|
|
end_point: self.end_point.into(),
|
|
}
|
|
}
|
|
}
|
|
|
|
impl From<ffi::TSRange> for Range {
|
|
fn from(range: ffi::TSRange) -> Self {
|
|
Self {
|
|
start_byte: range.start_byte as usize,
|
|
end_byte: range.end_byte as usize,
|
|
start_point: range.start_point.into(),
|
|
end_point: range.end_point.into(),
|
|
}
|
|
}
|
|
}
|
|
|
|
impl<'a> Into<ffi::TSInputEdit> for &'a InputEdit {
|
|
fn into(self) -> ffi::TSInputEdit {
|
|
ffi::TSInputEdit {
|
|
start_byte: self.start_byte as u32,
|
|
old_end_byte: self.old_end_byte as u32,
|
|
new_end_byte: self.new_end_byte as u32,
|
|
start_point: self.start_position.into(),
|
|
old_end_point: self.old_end_position.into(),
|
|
new_end_point: self.new_end_position.into(),
|
|
}
|
|
}
|
|
}
|
|
|
|
impl<'a> LossyUtf8<'a> {
|
|
pub fn new(bytes: &'a [u8]) -> Self {
|
|
LossyUtf8 {
|
|
bytes,
|
|
in_replacement: false,
|
|
}
|
|
}
|
|
}
|
|
|
|
impl<'a> Iterator for LossyUtf8<'a> {
|
|
type Item = &'a str;
|
|
|
|
fn next(&mut self) -> Option<&'a str> {
|
|
if self.bytes.is_empty() {
|
|
return None;
|
|
}
|
|
if self.in_replacement {
|
|
self.in_replacement = false;
|
|
return Some("\u{fffd}");
|
|
}
|
|
match std::str::from_utf8(self.bytes) {
|
|
Ok(valid) => {
|
|
self.bytes = &[];
|
|
Some(valid)
|
|
}
|
|
Err(error) => {
|
|
if let Some(error_len) = error.error_len() {
|
|
let error_start = error.valid_up_to();
|
|
if error_start > 0 {
|
|
let result =
|
|
unsafe { std::str::from_utf8_unchecked(&self.bytes[..error_start]) };
|
|
self.bytes = &self.bytes[(error_start + error_len)..];
|
|
self.in_replacement = true;
|
|
Some(result)
|
|
} else {
|
|
self.bytes = &self.bytes[error_len..];
|
|
Some("\u{fffd}")
|
|
}
|
|
} else {
|
|
None
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
fn predicate_error(row: usize, message: String) -> QueryError {
|
|
QueryError {
|
|
kind: QueryErrorKind::Predicate,
|
|
row,
|
|
column: 0,
|
|
offset: 0,
|
|
message,
|
|
}
|
|
}
|
|
|
|
impl fmt::Display for IncludedRangesError {
|
|
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
|
|
write!(f, "Incorrect range by index: {}", self.0)
|
|
}
|
|
}
|
|
|
|
impl fmt::Display for LanguageError {
|
|
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
|
|
write!(
|
|
f,
|
|
"Incompatible language version {}. Expected minimum {}, maximum {}",
|
|
self.version, MIN_COMPATIBLE_LANGUAGE_VERSION, LANGUAGE_VERSION,
|
|
)
|
|
}
|
|
}
|
|
|
|
impl fmt::Display for QueryError {
|
|
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
|
|
write!(
|
|
f,
|
|
"Query error at {}:{}. {}{}",
|
|
self.row + 1,
|
|
self.column + 1,
|
|
match self.kind {
|
|
QueryErrorKind::Field => "Invalid field name ",
|
|
QueryErrorKind::NodeType => "Invalid node type ",
|
|
QueryErrorKind::Capture => "Invalid capture name ",
|
|
QueryErrorKind::Predicate => "Invalid predicate: ",
|
|
QueryErrorKind::Structure => "Impossible pattern:\n",
|
|
QueryErrorKind::Syntax => "Invalid syntax:\n",
|
|
},
|
|
self.message
|
|
)
|
|
}
|
|
}
|
|
|
|
impl error::Error for IncludedRangesError {}
|
|
impl error::Error for LanguageError {}
|
|
impl error::Error for QueryError {}
|
|
|
|
unsafe impl Send for Language {}
|
|
unsafe impl Send for Parser {}
|
|
unsafe impl Send for Query {}
|
|
unsafe impl Send for QueryCursor {}
|
|
unsafe impl Send for Tree {}
|
|
unsafe impl Sync for Language {}
|
|
unsafe impl Sync for Parser {}
|
|
unsafe impl Sync for Query {}
|
|
unsafe impl Sync for QueryCursor {}
|
|
unsafe impl Sync for Tree {}
|