Unnamed repository; edit this file 'description' to name the repository.
Diffstat (limited to 'crates/proc-macro-srv/src/server_impl/token_stream.rs')
-rw-r--r--crates/proc-macro-srv/src/server_impl/token_stream.rs155
1 files changed, 155 insertions, 0 deletions
diff --git a/crates/proc-macro-srv/src/server_impl/token_stream.rs b/crates/proc-macro-srv/src/server_impl/token_stream.rs
new file mode 100644
index 0000000000..b1a448427c
--- /dev/null
+++ b/crates/proc-macro-srv/src/server_impl/token_stream.rs
@@ -0,0 +1,155 @@
+//! TokenStream implementation used by sysroot ABI
+
+use tt::TokenTree;
+
+#[derive(Debug, Clone)]
+pub struct TokenStream<S> {
+ pub(super) token_trees: Vec<TokenTree<S>>,
+}
+
+impl<S> Default for TokenStream<S> {
+ fn default() -> Self {
+ Self { token_trees: vec![] }
+ }
+}
+
+impl<S> TokenStream<S> {
+ pub(crate) fn new() -> Self {
+ TokenStream::default()
+ }
+
+ pub(crate) fn with_subtree(subtree: tt::Subtree<S>) -> Self {
+ if subtree.delimiter.kind != tt::DelimiterKind::Invisible {
+ TokenStream { token_trees: vec![TokenTree::Subtree(subtree)] }
+ } else {
+ TokenStream { token_trees: subtree.token_trees.into_vec() }
+ }
+ }
+
+ pub(crate) fn into_subtree(self, call_site: S) -> tt::Subtree<S>
+ where
+ S: Copy,
+ {
+ tt::Subtree {
+ delimiter: tt::Delimiter {
+ open: call_site,
+ close: call_site,
+ kind: tt::DelimiterKind::Invisible,
+ },
+ token_trees: self.token_trees.into_boxed_slice(),
+ }
+ }
+
+ pub(super) fn is_empty(&self) -> bool {
+ self.token_trees.is_empty()
+ }
+}
+
+/// Creates a token stream containing a single token tree.
+impl<S> From<TokenTree<S>> for TokenStream<S> {
+ fn from(tree: TokenTree<S>) -> TokenStream<S> {
+ TokenStream { token_trees: vec![tree] }
+ }
+}
+
+/// Collects a number of token trees into a single stream.
+impl<S> FromIterator<TokenTree<S>> for TokenStream<S> {
+ fn from_iter<I: IntoIterator<Item = TokenTree<S>>>(trees: I) -> Self {
+ trees.into_iter().map(TokenStream::from).collect()
+ }
+}
+
+/// A "flattening" operation on token streams, collects token trees
+/// from multiple token streams into a single stream.
+impl<S> FromIterator<TokenStream<S>> for TokenStream<S> {
+ fn from_iter<I: IntoIterator<Item = TokenStream<S>>>(streams: I) -> Self {
+ let mut builder = TokenStreamBuilder::new();
+ streams.into_iter().for_each(|stream| builder.push(stream));
+ builder.build()
+ }
+}
+
+impl<S> Extend<TokenTree<S>> for TokenStream<S> {
+ fn extend<I: IntoIterator<Item = TokenTree<S>>>(&mut self, trees: I) {
+ self.extend(trees.into_iter().map(TokenStream::from));
+ }
+}
+
+impl<S> Extend<TokenStream<S>> for TokenStream<S> {
+ fn extend<I: IntoIterator<Item = TokenStream<S>>>(&mut self, streams: I) {
+ for item in streams {
+ for tkn in item {
+ match tkn {
+ tt::TokenTree::Subtree(subtree)
+ if subtree.delimiter.kind == tt::DelimiterKind::Invisible =>
+ {
+ self.token_trees.extend(subtree.token_trees.into_vec().into_iter());
+ }
+ _ => {
+ self.token_trees.push(tkn);
+ }
+ }
+ }
+ }
+ }
+}
+
+pub(super) struct TokenStreamBuilder<S> {
+ acc: TokenStream<S>,
+}
+
+/// pub(super)lic implementation details for the `TokenStream` type, such as iterators.
+pub(super) mod token_stream {
+
+ use core::fmt;
+
+ use super::{TokenStream, TokenTree};
+
+ /// An iterator over `TokenStream`'s `TokenTree`s.
+ /// The iteration is "shallow", e.g., the iterator doesn't recurse into delimited groups,
+ /// and returns whole groups as token trees.
+ impl<S> IntoIterator for TokenStream<S> {
+ type Item = TokenTree<S>;
+ type IntoIter = std::vec::IntoIter<TokenTree<S>>;
+
+ fn into_iter(self) -> Self::IntoIter {
+ self.token_trees.into_iter()
+ }
+ }
+
+ /// Attempts to break the string into tokens and parse those tokens into a token stream.
+ /// May fail for a number of reasons, for example, if the string contains unbalanced delimiters
+ /// or characters not existing in the language.
+ /// All tokens in the parsed stream get `Span::call_site()` spans.
+ ///
+ /// NOTE: some errors may cause panics instead of returning `LexError`. We reserve the right to
+ /// change these errors into `LexError`s later.
+ impl<S: Copy + fmt::Debug> TokenStream<S> {
+ pub(crate) fn from_str(src: &str, call_site: S) -> Result<TokenStream<S>, String> {
+ let subtree =
+ mbe::parse_to_token_tree_static_span(call_site, src).ok_or("lexing error")?;
+
+ Ok(TokenStream::with_subtree(subtree))
+ }
+ }
+
+ impl<S> ToString for TokenStream<S> {
+ fn to_string(&self) -> String {
+ ::tt::pretty(&self.token_trees)
+ }
+ }
+}
+
+impl<S> TokenStreamBuilder<S> {
+ pub(super) fn new() -> TokenStreamBuilder<S> {
+ TokenStreamBuilder { acc: TokenStream::new() }
+ }
+
+ pub(super) fn push(&mut self, stream: TokenStream<S>) {
+ self.acc.extend(stream.into_iter())
+ }
+
+ pub(super) fn build(self) -> TokenStream<S> {
+ self.acc
+ }
+}