Unnamed repository; edit this file 'description' to name the repository.
Diffstat (limited to 'crates/hir-expand/src/db.rs')
-rw-r--r--crates/hir-expand/src/db.rs687
1 files changed, 380 insertions, 307 deletions
diff --git a/crates/hir-expand/src/db.rs b/crates/hir-expand/src/db.rs
index ff0d279d8c..d2c6559b06 100644
--- a/crates/hir-expand/src/db.rs
+++ b/crates/hir-expand/src/db.rs
@@ -1,22 +1,31 @@
//! Defines database & queries for macro expansion.
-use base_db::{salsa, CrateId, Edition, SourceDatabase};
+use base_db::{
+ salsa::{self, debug::DebugQueryTable},
+ span::SyntaxContextId,
+ CrateId, Edition, FileId, SourceDatabase,
+};
use either::Either;
use limit::Limit;
use mbe::{syntax_node_to_token_tree, ValueResult};
use rustc_hash::FxHashSet;
use syntax::{
ast::{self, HasAttrs, HasDocComments},
- AstNode, GreenNode, Parse, SyntaxError, SyntaxNode, SyntaxToken, T,
+ AstNode, Parse, SyntaxError, SyntaxNode, SyntaxToken, T,
};
use triomphe::Arc;
use crate::{
- ast_id_map::AstIdMap, builtin_attr_macro::pseudo_derive_attr_expansion,
- builtin_fn_macro::EagerExpander, fixup, hygiene::HygieneFrame, tt, AstId, BuiltinAttrExpander,
- BuiltinDeriveExpander, BuiltinFnLikeExpander, EagerCallInfo, ExpandError, ExpandResult,
- ExpandTo, HirFileId, HirFileIdRepr, MacroCallId, MacroCallKind, MacroCallLoc, MacroDefId,
- MacroDefKind, MacroFile, ProcMacroExpander,
+ ast_id_map::AstIdMap,
+ attrs::RawAttrs,
+ builtin_attr_macro::pseudo_derive_attr_expansion,
+ builtin_fn_macro::EagerExpander,
+ fixup::{self, SyntaxFixupUndoInfo},
+ hygiene::{apply_mark, SyntaxContextData, Transparency},
+ span::{RealSpanMap, SpanMap, SpanMapRef},
+ tt, AstId, BuiltinAttrExpander, BuiltinDeriveExpander, BuiltinFnLikeExpander, EagerCallInfo,
+ ExpandError, ExpandResult, ExpandTo, ExpansionSpanMap, HirFileId, HirFileIdRepr, MacroCallId,
+ MacroCallKind, MacroCallLoc, MacroDefId, MacroDefKind, MacroFileId, ProcMacroExpander,
};
/// Total limit on the number of tokens produced by any macro invocation.
@@ -30,32 +39,43 @@ static TOKEN_LIMIT: Limit = Limit::new(1_048_576);
#[derive(Debug, Clone, Eq, PartialEq)]
/// Old-style `macro_rules` or the new macros 2.0
pub struct DeclarativeMacroExpander {
- pub mac: mbe::DeclarativeMacro,
- pub def_site_token_map: mbe::TokenMap,
+ pub mac: mbe::DeclarativeMacro<base_db::span::SpanData>,
+ pub transparency: Transparency,
}
impl DeclarativeMacroExpander {
- pub fn expand(&self, tt: tt::Subtree) -> ExpandResult<tt::Subtree> {
+ pub fn expand(
+ &self,
+ db: &dyn ExpandDatabase,
+ tt: tt::Subtree,
+ call_id: MacroCallId,
+ ) -> ExpandResult<tt::Subtree> {
match self.mac.err() {
Some(e) => ExpandResult::new(
- tt::Subtree::empty(),
+ tt::Subtree::empty(tt::DelimSpan::DUMMY),
ExpandError::other(format!("invalid macro definition: {e}")),
),
- None => self.mac.expand(tt).map_err(Into::into),
+ None => self
+ .mac
+ .expand(&tt, |s| s.ctx = apply_mark(db, s.ctx, call_id, self.transparency))
+ .map_err(Into::into),
}
}
- pub fn map_id_down(&self, token_id: tt::TokenId) -> tt::TokenId {
- self.mac.map_id_down(token_id)
- }
-
- pub fn map_id_up(&self, token_id: tt::TokenId) -> (tt::TokenId, mbe::Origin) {
- self.mac.map_id_up(token_id)
+ pub fn expand_unhygienic(&self, tt: tt::Subtree) -> ExpandResult<tt::Subtree> {
+ match self.mac.err() {
+ Some(e) => ExpandResult::new(
+ tt::Subtree::empty(tt::DelimSpan::DUMMY),
+ ExpandError::other(format!("invalid macro definition: {e}")),
+ ),
+ None => self.mac.expand(&tt, |_| ()).map_err(Into::into),
+ }
}
}
#[derive(Debug, Clone, Eq, PartialEq)]
pub enum TokenExpander {
+ /// Old-style `macro_rules` or the new macros 2.0
DeclarativeMacro(Arc<DeclarativeMacroExpander>),
/// Stuff like `line!` and `file!`.
BuiltIn(BuiltinFnLikeExpander),
@@ -69,31 +89,6 @@ pub enum TokenExpander {
ProcMacro(ProcMacroExpander),
}
-// FIXME: Get rid of these methods
-impl TokenExpander {
- pub(crate) fn map_id_down(&self, id: tt::TokenId) -> tt::TokenId {
- match self {
- TokenExpander::DeclarativeMacro(expander) => expander.map_id_down(id),
- TokenExpander::BuiltIn(..)
- | TokenExpander::BuiltInEager(..)
- | TokenExpander::BuiltInAttr(..)
- | TokenExpander::BuiltInDerive(..)
- | TokenExpander::ProcMacro(..) => id,
- }
- }
-
- pub(crate) fn map_id_up(&self, id: tt::TokenId) -> (tt::TokenId, mbe::Origin) {
- match self {
- TokenExpander::DeclarativeMacro(expander) => expander.map_id_up(id),
- TokenExpander::BuiltIn(..)
- | TokenExpander::BuiltInEager(..)
- | TokenExpander::BuiltInAttr(..)
- | TokenExpander::BuiltInDerive(..)
- | TokenExpander::ProcMacro(..) => (id, mbe::Origin::Call),
- }
- }
-}
-
#[salsa::query_group(ExpandDatabaseStorage)]
pub trait ExpandDatabase: SourceDatabase {
fn ast_id_map(&self, file_id: HirFileId) -> Arc<AstIdMap>;
@@ -108,8 +103,12 @@ pub trait ExpandDatabase: SourceDatabase {
// This query is LRU cached
fn parse_macro_expansion(
&self,
- macro_file: MacroFile,
- ) -> ExpandResult<(Parse<SyntaxNode>, Arc<mbe::TokenMap>)>;
+ macro_file: MacroFileId,
+ ) -> ExpandResult<(Parse<SyntaxNode>, Arc<ExpansionSpanMap>)>;
+ #[salsa::transparent]
+ fn span_map(&self, file_id: HirFileId) -> SpanMap;
+
+ fn real_span_map(&self, file_id: FileId) -> Arc<RealSpanMap>;
/// Macro ids. That's probably the tricksiest bit in rust-analyzer, and the
/// reason why we use salsa at all.
@@ -118,23 +117,21 @@ pub trait ExpandDatabase: SourceDatabase {
/// to be incremental.
#[salsa::interned]
fn intern_macro_call(&self, macro_call: MacroCallLoc) -> MacroCallId;
+ #[salsa::interned]
+ fn intern_syntax_context(&self, ctx: SyntaxContextData) -> SyntaxContextId;
- /// Lowers syntactic macro call to a token tree representation.
#[salsa::transparent]
- fn macro_arg(
- &self,
- id: MacroCallId,
- ) -> ValueResult<
- Option<Arc<(tt::Subtree, mbe::TokenMap, fixup::SyntaxFixupUndoInfo)>>,
- Arc<Box<[SyntaxError]>>,
- >;
- /// Extracts syntax node, corresponding to a macro call. That's a firewall
+ fn setup_syntax_context_root(&self) -> ();
+ #[salsa::transparent]
+ fn dump_syntax_contexts(&self) -> String;
+
+ /// Lowers syntactic macro call to a token tree representation. That's a firewall
/// query, only typing in the macro call itself changes the returned
/// subtree.
- fn macro_arg_node(
+ fn macro_arg(
&self,
id: MacroCallId,
- ) -> ValueResult<Option<GreenNode>, Arc<Box<[SyntaxError]>>>;
+ ) -> ValueResult<Option<(Arc<tt::Subtree>, SyntaxFixupUndoInfo)>, Arc<Box<[SyntaxError]>>>;
/// Fetches the expander for this macro.
#[salsa::transparent]
fn macro_expander(&self, id: MacroDefId) -> TokenExpander;
@@ -144,18 +141,6 @@ pub trait ExpandDatabase: SourceDatabase {
def_crate: CrateId,
id: AstId<ast::Macro>,
) -> Arc<DeclarativeMacroExpander>;
-
- /// Expand macro call to a token tree.
- // This query is LRU cached
- fn macro_expand(&self, macro_call: MacroCallId) -> ExpandResult<Arc<tt::Subtree>>;
- #[salsa::invoke(crate::builtin_fn_macro::include_arg_to_tt)]
- fn include_expand(
- &self,
- arg_id: MacroCallId,
- ) -> Result<
- (triomphe::Arc<(::tt::Subtree<::tt::TokenId>, mbe::TokenMap)>, base_db::FileId),
- ExpandError,
- >;
/// Special case of the previous query for procedural macros. We can't LRU
/// proc macros, since they are not deterministic in general, and
/// non-determinism breaks salsa in a very, very, very bad way.
@@ -166,8 +151,20 @@ pub trait ExpandDatabase: SourceDatabase {
&self,
macro_call: MacroCallId,
) -> ExpandResult<Box<[SyntaxError]>>;
+}
- fn hygiene_frame(&self, file_id: HirFileId) -> Arc<HygieneFrame>;
+#[inline]
+pub fn span_map(db: &dyn ExpandDatabase, file_id: HirFileId) -> SpanMap {
+ match file_id.repr() {
+ HirFileIdRepr::FileId(file_id) => SpanMap::RealSpanMap(db.real_span_map(file_id)),
+ HirFileIdRepr::MacroFile(m) => {
+ SpanMap::ExpansionSpanMap(db.parse_macro_expansion(m).value.1)
+ }
+ }
+}
+
+pub fn real_span_map(db: &dyn ExpandDatabase, file_id: FileId) -> Arc<RealSpanMap> {
+ Arc::new(RealSpanMap::from_file(db, file_id))
}
/// This expands the given macro call, but with different arguments. This is
@@ -181,21 +178,36 @@ pub fn expand_speculative(
token_to_map: SyntaxToken,
) -> Option<(SyntaxNode, SyntaxToken)> {
let loc = db.lookup_intern_macro_call(actual_macro_call);
- let token_range = token_to_map.text_range();
+
+ let span_map = RealSpanMap::absolute(FileId::BOGUS);
+ let span_map = SpanMapRef::RealSpanMap(&span_map);
// Build the subtree and token mapping for the speculative args
- let censor = censor_for_macro_input(&loc, speculative_args);
- let mut fixups = fixup::fixup_syntax(speculative_args);
- fixups.replace.extend(censor.into_iter().map(|node| (node.into(), Vec::new())));
- let (mut tt, spec_args_tmap, _) = mbe::syntax_node_to_token_tree_with_modifications(
- speculative_args,
- fixups.token_map,
- fixups.next_id,
- fixups.replace,
- fixups.append,
- );
+ let (mut tt, undo_info) = match loc.kind {
+ MacroCallKind::FnLike { .. } => {
+ (mbe::syntax_node_to_token_tree(speculative_args, span_map), SyntaxFixupUndoInfo::NONE)
+ }
+ MacroCallKind::Derive { .. } | MacroCallKind::Attr { .. } => {
+ let censor = censor_for_macro_input(&loc, speculative_args);
+ let mut fixups = fixup::fixup_syntax(span_map, speculative_args);
+ fixups.append.retain(|it, _| match it {
+ syntax::NodeOrToken::Node(it) => !censor.contains(it),
+ syntax::NodeOrToken::Token(_) => true,
+ });
+ fixups.remove.extend(censor);
+ (
+ mbe::syntax_node_to_token_tree_modified(
+ speculative_args,
+ span_map,
+ fixups.append,
+ fixups.remove,
+ ),
+ fixups.undo_info,
+ )
+ }
+ };
- let (attr_arg, token_id) = match loc.kind {
+ let attr_arg = match loc.kind {
MacroCallKind::Attr { invoc_attr_index, .. } => {
let attr = if loc.def.is_attribute_derive() {
// for pseudo-derive expansion we actually pass the attribute itself only
@@ -210,59 +222,45 @@ pub fn expand_speculative(
}?;
match attr.token_tree() {
Some(token_tree) => {
- let (mut tree, map) = syntax_node_to_token_tree(attr.token_tree()?.syntax());
- tree.delimiter = tt::Delimiter::unspecified();
-
- let shift = mbe::Shift::new(&tt);
- shift.shift_all(&mut tree);
-
- let token_id = if token_tree.syntax().text_range().contains_range(token_range) {
- let attr_input_start =
- token_tree.left_delimiter_token()?.text_range().start();
- let range = token_range.checked_sub(attr_input_start)?;
- let token_id = shift.shift(map.token_by_range(range)?);
- Some(token_id)
- } else {
- None
- };
- (Some(tree), token_id)
- }
- _ => (None, None),
- }
- }
- _ => (None, None),
- };
- let token_id = match token_id {
- Some(token_id) => token_id,
- // token wasn't inside an attribute input so it has to be in the general macro input
- None => {
- let range = token_range.checked_sub(speculative_args.text_range().start())?;
- let token_id = spec_args_tmap.token_by_range(range)?;
- match loc.def.kind {
- MacroDefKind::Declarative(it) => {
- db.decl_macro_expander(loc.krate, it).map_id_down(token_id)
+ let mut tree = syntax_node_to_token_tree(token_tree.syntax(), span_map);
+ tree.delimiter = tt::Delimiter::DUMMY_INVISIBLE;
+
+ Some(tree)
}
- _ => token_id,
+ _ => None,
}
}
+ _ => None,
};
// Do the actual expansion, we need to directly expand the proc macro due to the attribute args
// Otherwise the expand query will fetch the non speculative attribute args and pass those instead.
let mut speculative_expansion = match loc.def.kind {
MacroDefKind::ProcMacro(expander, ..) => {
- tt.delimiter = tt::Delimiter::unspecified();
- expander.expand(db, loc.def.krate, loc.krate, &tt, attr_arg.as_ref())
+ tt.delimiter = tt::Delimiter::DUMMY_INVISIBLE;
+ let call_site = loc.span(db);
+ expander.expand(
+ db,
+ loc.def.krate,
+ loc.krate,
+ &tt,
+ attr_arg.as_ref(),
+ call_site,
+ call_site,
+ call_site,
+ )
}
MacroDefKind::BuiltInAttr(BuiltinAttrExpander::Derive, _) => {
- pseudo_derive_attr_expansion(&tt, attr_arg.as_ref()?)
+ pseudo_derive_attr_expansion(&tt, attr_arg.as_ref()?, loc.call_site)
}
MacroDefKind::BuiltInDerive(expander, ..) => {
// this cast is a bit sus, can we avoid losing the typedness here?
let adt = ast::Adt::cast(speculative_args.clone()).unwrap();
- expander.expand(db, actual_macro_call, &adt, &spec_args_tmap)
+ expander.expand(db, actual_macro_call, &adt, span_map)
+ }
+ MacroDefKind::Declarative(it) => {
+ db.decl_macro_expander(loc.krate, it).expand_unhygienic(tt)
}
- MacroDefKind::Declarative(it) => db.decl_macro_expander(loc.krate, it).expand(tt),
MacroDefKind::BuiltIn(it, _) => it.expand(db, actual_macro_call, &tt).map_err(Into::into),
MacroDefKind::BuiltInEager(it, _) => {
it.expand(db, actual_macro_call, &tt).map_err(Into::into)
@@ -270,13 +268,14 @@ pub fn expand_speculative(
MacroDefKind::BuiltInAttr(it, _) => it.expand(db, actual_macro_call, &tt),
};
- let expand_to = macro_expand_to(db, actual_macro_call);
- fixup::reverse_fixups(&mut speculative_expansion.value, &spec_args_tmap, &fixups.undo_info);
+ let expand_to = loc.expand_to();
+
+ fixup::reverse_fixups(&mut speculative_expansion.value, &undo_info);
let (node, rev_tmap) = token_tree_to_syntax_node(&speculative_expansion.value, expand_to);
let syntax_node = node.syntax_node();
let token = rev_tmap
- .ranges_by_token(token_id, token_to_map.kind())
+ .ranges_with_span(span_map.span_for_range(token_to_map.text_range()))
.filter_map(|range| syntax_node.covering_element(range).into_token())
.min_by_key(|t| {
// prefer tokens of the same kind and text
@@ -293,7 +292,7 @@ fn ast_id_map(db: &dyn ExpandDatabase, file_id: HirFileId) -> Arc<AstIdMap> {
fn parse_or_expand(db: &dyn ExpandDatabase, file_id: HirFileId) -> SyntaxNode {
match file_id.repr() {
- HirFileIdRepr::FileId(file_id) => db.parse(file_id).tree().syntax().clone(),
+ HirFileIdRepr::FileId(file_id) => db.parse(file_id).syntax_node(),
HirFileIdRepr::MacroFile(macro_file) => {
db.parse_macro_expansion(macro_file).value.0.syntax_node()
}
@@ -312,17 +311,16 @@ fn parse_or_expand_with_err(
}
}
+// FIXME: We should verify that the parsed node is one of the many macro node variants we expect
+// instead of having it be untyped
fn parse_macro_expansion(
db: &dyn ExpandDatabase,
- macro_file: MacroFile,
-) -> ExpandResult<(Parse<SyntaxNode>, Arc<mbe::TokenMap>)> {
+ macro_file: MacroFileId,
+) -> ExpandResult<(Parse<SyntaxNode>, Arc<ExpansionSpanMap>)> {
let _p = profile::span("parse_macro_expansion");
- let mbe::ValueResult { value: tt, err } = db.macro_expand(macro_file.macro_call_id);
-
- let expand_to = macro_expand_to(db, macro_file.macro_call_id);
-
- tracing::debug!("expanded = {}", tt.as_debug_string());
- tracing::debug!("kind = {:?}", expand_to);
+ let loc = db.lookup_intern_macro_call(macro_file.macro_call_id);
+ let expand_to = loc.expand_to();
+ let mbe::ValueResult { value: tt, err } = macro_expand(db, macro_file.macro_call_id, loc);
let (parse, rev_token_map) = token_tree_to_syntax_node(&tt, expand_to);
@@ -333,51 +331,129 @@ fn parse_macro_expansion_error(
db: &dyn ExpandDatabase,
macro_call_id: MacroCallId,
) -> ExpandResult<Box<[SyntaxError]>> {
- db.parse_macro_expansion(MacroFile { macro_call_id })
+ db.parse_macro_expansion(MacroFileId { macro_call_id })
.map(|it| it.0.errors().to_vec().into_boxed_slice())
}
+fn parse_with_map(db: &dyn ExpandDatabase, file_id: HirFileId) -> (Parse<SyntaxNode>, SpanMap) {
+ match file_id.repr() {
+ HirFileIdRepr::FileId(file_id) => {
+ (db.parse(file_id).to_syntax(), SpanMap::RealSpanMap(db.real_span_map(file_id)))
+ }
+ HirFileIdRepr::MacroFile(macro_file) => {
+ let (parse, map) = db.parse_macro_expansion(macro_file).value;
+ (parse, SpanMap::ExpansionSpanMap(map))
+ }
+ }
+}
+
fn macro_arg(
db: &dyn ExpandDatabase,
id: MacroCallId,
-) -> ValueResult<
- Option<Arc<(tt::Subtree, mbe::TokenMap, fixup::SyntaxFixupUndoInfo)>>,
- Arc<Box<[SyntaxError]>>,
-> {
- let loc = db.lookup_intern_macro_call(id);
-
- if let Some(EagerCallInfo { arg, arg_id: _, error: _ }) = loc.eager.as_deref() {
- return ValueResult::ok(Some(Arc::new((arg.0.clone(), arg.1.clone(), Default::default()))));
- }
-
- let ValueResult { value, err } = db.macro_arg_node(id);
- let Some(arg) = value else {
- return ValueResult { value: None, err };
+ // FIXME: consider the following by putting fixup info into eager call info args
+ // ) -> ValueResult<Option<Arc<(tt::Subtree, SyntaxFixupUndoInfo)>>, Arc<Box<[SyntaxError]>>> {
+) -> ValueResult<Option<(Arc<tt::Subtree>, SyntaxFixupUndoInfo)>, Arc<Box<[SyntaxError]>>> {
+ let mismatched_delimiters = |arg: &SyntaxNode| {
+ let first = arg.first_child_or_token().map_or(T![.], |it| it.kind());
+ let last = arg.last_child_or_token().map_or(T![.], |it| it.kind());
+ let well_formed_tt =
+ matches!((first, last), (T!['('], T![')']) | (T!['['], T![']']) | (T!['{'], T!['}']));
+ if !well_formed_tt {
+ // Don't expand malformed (unbalanced) macro invocations. This is
+ // less than ideal, but trying to expand unbalanced macro calls
+ // sometimes produces pathological, deeply nested code which breaks
+ // all kinds of things.
+ //
+ // Some day, we'll have explicit recursion counters for all
+ // recursive things, at which point this code might be removed.
+ cov_mark::hit!(issue9358_bad_macro_stack_overflow);
+ Some(Arc::new(Box::new([SyntaxError::new(
+ "unbalanced token tree".to_owned(),
+ arg.text_range(),
+ )]) as Box<[_]>))
+ } else {
+ None
+ }
};
+ let loc = db.lookup_intern_macro_call(id);
+ if let Some(EagerCallInfo { arg, .. }) = matches!(loc.def.kind, MacroDefKind::BuiltInEager(..))
+ .then(|| loc.eager.as_deref())
+ .flatten()
+ {
+ ValueResult::ok(Some((arg.clone(), SyntaxFixupUndoInfo::NONE)))
+ } else {
+ let (parse, map) = parse_with_map(db, loc.kind.file_id());
+ let root = parse.syntax_node();
+
+ let syntax = match loc.kind {
+ MacroCallKind::FnLike { ast_id, .. } => {
+ let node = &ast_id.to_ptr(db).to_node(&root);
+ let offset = node.syntax().text_range().start();
+ match node.token_tree() {
+ Some(tt) => {
+ let tt = tt.syntax();
+ if let Some(e) = mismatched_delimiters(tt) {
+ return ValueResult::only_err(e);
+ }
+ tt.clone()
+ }
+ None => {
+ return ValueResult::only_err(Arc::new(Box::new([
+ SyntaxError::new_at_offset("missing token tree".to_owned(), offset),
+ ])));
+ }
+ }
+ }
+ MacroCallKind::Derive { ast_id, .. } => {
+ ast_id.to_ptr(db).to_node(&root).syntax().clone()
+ }
+ MacroCallKind::Attr { ast_id, .. } => ast_id.to_ptr(db).to_node(&root).syntax().clone(),
+ };
+ let (mut tt, undo_info) = match loc.kind {
+ MacroCallKind::FnLike { .. } => {
+ (mbe::syntax_node_to_token_tree(&syntax, map.as_ref()), SyntaxFixupUndoInfo::NONE)
+ }
+ MacroCallKind::Derive { .. } | MacroCallKind::Attr { .. } => {
+ let censor = censor_for_macro_input(&loc, &syntax);
+ let mut fixups = fixup::fixup_syntax(map.as_ref(), &syntax);
+ fixups.append.retain(|it, _| match it {
+ syntax::NodeOrToken::Node(it) => !censor.contains(it),
+ syntax::NodeOrToken::Token(_) => true,
+ });
+ fixups.remove.extend(censor);
+ (
+ mbe::syntax_node_to_token_tree_modified(
+ &syntax,
+ map,
+ fixups.append,
+ fixups.remove,
+ ),
+ fixups.undo_info,
+ )
+ }
+ };
- let node = SyntaxNode::new_root(arg);
- let censor = censor_for_macro_input(&loc, &node);
- let mut fixups = fixup::fixup_syntax(&node);
- fixups.replace.extend(censor.into_iter().map(|node| (node.into(), Vec::new())));
- let (mut tt, tmap, _) = mbe::syntax_node_to_token_tree_with_modifications(
- &node,
- fixups.token_map,
- fixups.next_id,
- fixups.replace,
- fixups.append,
- );
+ if loc.def.is_proc_macro() {
+ // proc macros expect their inputs without parentheses, MBEs expect it with them included
+ tt.delimiter = tt::Delimiter::DUMMY_INVISIBLE;
+ }
- if loc.def.is_proc_macro() {
- // proc macros expect their inputs without parentheses, MBEs expect it with them included
- tt.delimiter = tt::Delimiter::unspecified();
- }
- let val = Some(Arc::new((tt, tmap, fixups.undo_info)));
- match err {
- Some(err) => ValueResult::new(val, err),
- None => ValueResult::ok(val),
+ if matches!(loc.def.kind, MacroDefKind::BuiltInEager(..)) {
+ match parse.errors() {
+ [] => ValueResult::ok(Some((Arc::new(tt), undo_info))),
+ errors => ValueResult::new(
+ Some((Arc::new(tt), undo_info)),
+ // Box::<[_]>::from(res.errors()), not stable yet
+ Arc::new(errors.to_vec().into_boxed_slice()),
+ ),
+ }
+ } else {
+ ValueResult::ok(Some((Arc::new(tt), undo_info)))
+ }
}
}
+// FIXME: Censoring info should be calculated by the caller! Namely by name resolution
/// Certain macro calls expect some nodes in the input to be preprocessed away, namely:
/// - derives expect all `#[derive(..)]` invocations up to the currently invoked one to be stripped
/// - attributes expect the invoking attribute to be stripped
@@ -417,103 +493,67 @@ fn censor_for_macro_input(loc: &MacroCallLoc, node: &SyntaxNode) -> FxHashSet<Sy
.unwrap_or_default()
}
-fn macro_arg_node(
- db: &dyn ExpandDatabase,
- id: MacroCallId,
-) -> ValueResult<Option<GreenNode>, Arc<Box<[SyntaxError]>>> {
- let err = || -> Arc<Box<[_]>> {
- Arc::new(Box::new([SyntaxError::new_at_offset(
- "invalid macro call".to_owned(),
- syntax::TextSize::from(0),
- )]))
- };
- let loc = db.lookup_intern_macro_call(id);
- let arg = if let MacroDefKind::BuiltInEager(..) = loc.def.kind {
- let res = if let Some(EagerCallInfo { arg, .. }) = loc.eager.as_deref() {
- Some(mbe::token_tree_to_syntax_node(&arg.0, mbe::TopEntryPoint::MacroEagerInput).0)
- } else {
- loc.kind
- .arg(db)
- .and_then(|arg| ast::TokenTree::cast(arg.value))
- .map(|tt| tt.reparse_as_comma_separated_expr().to_syntax())
- };
- match res {
- Some(res) if res.errors().is_empty() => res.syntax_node(),
- Some(res) => {
- return ValueResult::new(
- Some(res.syntax_node().green().into()),
- // Box::<[_]>::from(res.errors()), not stable yet
- Arc::new(res.errors().to_vec().into_boxed_slice()),
- );
- }
- None => return ValueResult::only_err(err()),
- }
- } else {
- match loc.kind.arg(db) {
- Some(res) => res.value,
- None => return ValueResult::only_err(err()),
- }
- };
- if matches!(loc.kind, MacroCallKind::FnLike { .. }) {
- let first = arg.first_child_or_token().map_or(T![.], |it| it.kind());
- let last = arg.last_child_or_token().map_or(T![.], |it| it.kind());
- let well_formed_tt =
- matches!((first, last), (T!['('], T![')']) | (T!['['], T![']']) | (T!['{'], T!['}']));
- if !well_formed_tt {
- // Don't expand malformed (unbalanced) macro invocations. This is
- // less than ideal, but trying to expand unbalanced macro calls
- // sometimes produces pathological, deeply nested code which breaks
- // all kinds of things.
- //
- // Some day, we'll have explicit recursion counters for all
- // recursive things, at which point this code might be removed.
- cov_mark::hit!(issue9358_bad_macro_stack_overflow);
- return ValueResult::only_err(Arc::new(Box::new([SyntaxError::new(
- "unbalanced token tree".to_owned(),
- arg.text_range(),
- )])));
- }
- }
- ValueResult::ok(Some(arg.green().into()))
-}
-
fn decl_macro_expander(
db: &dyn ExpandDatabase,
def_crate: CrateId,
id: AstId<ast::Macro>,
) -> Arc<DeclarativeMacroExpander> {
let is_2021 = db.crate_graph()[def_crate].edition >= Edition::Edition2021;
- let (mac, def_site_token_map) = match id.to_node(db) {
- ast::Macro::MacroRules(macro_rules) => match macro_rules.token_tree() {
- Some(arg) => {
- let (tt, def_site_token_map) = mbe::syntax_node_to_token_tree(arg.syntax());
- let mac = mbe::DeclarativeMacro::parse_macro_rules(&tt, is_2021);
- (mac, def_site_token_map)
- }
- None => (
- mbe::DeclarativeMacro::from_err(
+ let (root, map) = parse_with_map(db, id.file_id);
+ let root = root.syntax_node();
+
+ let transparency = |node| {
+ // ... would be nice to have the item tree here
+ let attrs = RawAttrs::new(db, node, map.as_ref()).filter(db, def_crate);
+ match &*attrs
+ .iter()
+ .find(|it| {
+ it.path.as_ident().and_then(|it| it.as_str()) == Some("rustc_macro_transparency")
+ })?
+ .token_tree_value()?
+ .token_trees
+ {
+ [tt::TokenTree::Leaf(tt::Leaf::Ident(i)), ..] => match &*i.text {
+ "transparent" => Some(Transparency::Transparent),
+ "semitransparent" => Some(Transparency::SemiTransparent),
+ "opaque" => Some(Transparency::Opaque),
+ _ => None,
+ },
+ _ => None,
+ }
+ };
+
+ let (mac, transparency) = match id.to_ptr(db).to_node(&root) {
+ ast::Macro::MacroRules(macro_rules) => (
+ match macro_rules.token_tree() {
+ Some(arg) => {
+ let tt = mbe::syntax_node_to_token_tree(arg.syntax(), map.as_ref());
+ let mac = mbe::DeclarativeMacro::parse_macro_rules(&tt, is_2021);
+ mac
+ }
+ None => mbe::DeclarativeMacro::from_err(
mbe::ParseError::Expected("expected a token tree".into()),
is_2021,
),
- Default::default(),
- ),
- },
- ast::Macro::MacroDef(macro_def) => match macro_def.body() {
- Some(arg) => {
- let (tt, def_site_token_map) = mbe::syntax_node_to_token_tree(arg.syntax());
- let mac = mbe::DeclarativeMacro::parse_macro2(&tt, is_2021);
- (mac, def_site_token_map)
- }
- None => (
- mbe::DeclarativeMacro::from_err(
+ },
+ transparency(&macro_rules).unwrap_or(Transparency::SemiTransparent),
+ ),
+ ast::Macro::MacroDef(macro_def) => (
+ match macro_def.body() {
+ Some(arg) => {
+ let tt = mbe::syntax_node_to_token_tree(arg.syntax(), map.as_ref());
+ let mac = mbe::DeclarativeMacro::parse_macro2(&tt, is_2021);
+ mac
+ }
+ None => mbe::DeclarativeMacro::from_err(
mbe::ParseError::Expected("expected a token tree".into()),
is_2021,
),
- Default::default(),
- ),
- },
+ },
+ transparency(&macro_def).unwrap_or(Transparency::Opaque),
+ ),
};
- Arc::new(DeclarativeMacroExpander { mac, def_site_token_map })
+ Arc::new(DeclarativeMacroExpander { mac, transparency })
}
fn macro_expander(db: &dyn ExpandDatabase, id: MacroDefId) -> TokenExpander {
@@ -529,39 +569,31 @@ fn macro_expander(db: &dyn ExpandDatabase, id: MacroDefId) -> TokenExpander {
}
}
-fn macro_expand(db: &dyn ExpandDatabase, id: MacroCallId) -> ExpandResult<Arc<tt::Subtree>> {
+fn macro_expand(
+ db: &dyn ExpandDatabase,
+ macro_call_id: MacroCallId,
+ loc: MacroCallLoc,
+) -> ExpandResult<Arc<tt::Subtree>> {
let _p = profile::span("macro_expand");
- let loc = db.lookup_intern_macro_call(id);
let ExpandResult { value: tt, mut err } = match loc.def.kind {
- MacroDefKind::ProcMacro(..) => return db.expand_proc_macro(id),
+ MacroDefKind::ProcMacro(..) => return db.expand_proc_macro(macro_call_id),
MacroDefKind::BuiltInDerive(expander, ..) => {
- let arg = db.macro_arg_node(id).value.unwrap();
-
- let node = SyntaxNode::new_root(arg);
- let censor = censor_for_macro_input(&loc, &node);
- let mut fixups = fixup::fixup_syntax(&node);
- fixups.replace.extend(censor.into_iter().map(|node| (node.into(), Vec::new())));
- let (tmap, _) = mbe::syntax_node_to_token_map_with_modifications(
- &node,
- fixups.token_map,
- fixups.next_id,
- fixups.replace,
- fixups.append,
- );
-
- // this cast is a bit sus, can we avoid losing the typedness here?
- let adt = ast::Adt::cast(node).unwrap();
- let mut res = expander.expand(db, id, &adt, &tmap);
- fixup::reverse_fixups(&mut res.value, &tmap, &fixups.undo_info);
- res
+ let (root, map) = parse_with_map(db, loc.kind.file_id());
+ let root = root.syntax_node();
+ let MacroCallKind::Derive { ast_id, .. } = loc.kind else { unreachable!() };
+ let node = ast_id.to_ptr(db).to_node(&root);
+
+ // FIXME: Use censoring
+ let _censor = censor_for_macro_input(&loc, node.syntax());
+ expander.expand(db, macro_call_id, &node, map.as_ref())
}
_ => {
- let ValueResult { value, err } = db.macro_arg(id);
- let Some(macro_arg) = value else {
+ let ValueResult { value, err } = db.macro_arg(macro_call_id);
+ let Some((macro_arg, undo_info)) = value else {
return ExpandResult {
value: Arc::new(tt::Subtree {
- delimiter: tt::Delimiter::UNSPECIFIED,
+ delimiter: tt::Delimiter::DUMMY_INVISIBLE,
token_trees: Vec::new(),
}),
// FIXME: We should make sure to enforce an invariant that invalid macro
@@ -570,12 +602,14 @@ fn macro_expand(db: &dyn ExpandDatabase, id: MacroCallId) -> ExpandResult<Arc<tt
};
};
- let (arg, arg_tm, undo_info) = &*macro_arg;
- let mut res = match loc.def.kind {
+ let arg = &*macro_arg;
+ match loc.def.kind {
MacroDefKind::Declarative(id) => {
- db.decl_macro_expander(loc.def.krate, id).expand(arg.clone())
+ db.decl_macro_expander(loc.def.krate, id).expand(db, arg.clone(), macro_call_id)
+ }
+ MacroDefKind::BuiltIn(it, _) => {
+ it.expand(db, macro_call_id, &arg).map_err(Into::into)
}
- MacroDefKind::BuiltIn(it, _) => it.expand(db, id, &arg).map_err(Into::into),
// This might look a bit odd, but we do not expand the inputs to eager macros here.
// Eager macros inputs are expanded, well, eagerly when we collect the macro calls.
// That kind of expansion uses the ast id map of an eager macros input though which goes through
@@ -583,11 +617,8 @@ fn macro_expand(db: &dyn ExpandDatabase, id: MacroCallId) -> ExpandResult<Arc<tt
// will end up going through here again, whereas we want to just want to inspect the raw input.
// As such we just return the input subtree here.
MacroDefKind::BuiltInEager(..) if loc.eager.is_none() => {
- let mut arg = arg.clone();
- fixup::reverse_fixups(&mut arg, arg_tm, undo_info);
-
return ExpandResult {
- value: Arc::new(arg),
+ value: macro_arg.clone(),
err: err.map(|err| {
let mut buf = String::new();
for err in &**err {
@@ -600,12 +631,16 @@ fn macro_expand(db: &dyn ExpandDatabase, id: MacroCallId) -> ExpandResult<Arc<tt
}),
};
}
- MacroDefKind::BuiltInEager(it, _) => it.expand(db, id, &arg).map_err(Into::into),
- MacroDefKind::BuiltInAttr(it, _) => it.expand(db, id, &arg),
+ MacroDefKind::BuiltInEager(it, _) => {
+ it.expand(db, macro_call_id, &arg).map_err(Into::into)
+ }
+ MacroDefKind::BuiltInAttr(it, _) => {
+ let mut res = it.expand(db, macro_call_id, &arg);
+ fixup::reverse_fixups(&mut res.value, &undo_info);
+ res
+ }
_ => unreachable!(),
- };
- fixup::reverse_fixups(&mut res.value, arg_tm, undo_info);
- res
+ }
}
};
@@ -627,10 +662,10 @@ fn macro_expand(db: &dyn ExpandDatabase, id: MacroCallId) -> ExpandResult<Arc<tt
fn expand_proc_macro(db: &dyn ExpandDatabase, id: MacroCallId) -> ExpandResult<Arc<tt::Subtree>> {
let loc = db.lookup_intern_macro_call(id);
- let Some(macro_arg) = db.macro_arg(id).value else {
+ let Some((macro_arg, undo_info)) = db.macro_arg(id).value else {
return ExpandResult {
value: Arc::new(tt::Subtree {
- delimiter: tt::Delimiter::UNSPECIFIED,
+ delimiter: tt::Delimiter::DUMMY_INVISIBLE,
token_trees: Vec::new(),
}),
// FIXME: We should make sure to enforce an invariant that invalid macro
@@ -639,47 +674,44 @@ fn expand_proc_macro(db: &dyn ExpandDatabase, id: MacroCallId) -> ExpandResult<A
};
};
- let (arg_tt, arg_tm, undo_info) = &*macro_arg;
-
let expander = match loc.def.kind {
MacroDefKind::ProcMacro(expander, ..) => expander,
_ => unreachable!(),
};
let attr_arg = match &loc.kind {
- MacroCallKind::Attr { attr_args, .. } => {
- let mut attr_args = attr_args.0.clone();
- mbe::Shift::new(arg_tt).shift_all(&mut attr_args);
- Some(attr_args)
- }
+ MacroCallKind::Attr { attr_args: Some(attr_args), .. } => Some(&**attr_args),
_ => None,
};
- let ExpandResult { value: mut tt, err } =
- expander.expand(db, loc.def.krate, loc.krate, arg_tt, attr_arg.as_ref());
+ let call_site = loc.span(db);
+ let ExpandResult { value: mut tt, err } = expander.expand(
+ db,
+ loc.def.krate,
+ loc.krate,
+ &macro_arg,
+ attr_arg,
+ // FIXME
+ call_site,
+ call_site,
+ // FIXME
+ call_site,
+ );
// Set a hard limit for the expanded tt
if let Err(value) = check_tt_count(&tt) {
return value;
}
- fixup::reverse_fixups(&mut tt, arg_tm, undo_info);
+ fixup::reverse_fixups(&mut tt, &undo_info);
ExpandResult { value: Arc::new(tt), err }
}
-fn hygiene_frame(db: &dyn ExpandDatabase, file_id: HirFileId) -> Arc<HygieneFrame> {
- Arc::new(HygieneFrame::new(db, file_id))
-}
-
-fn macro_expand_to(db: &dyn ExpandDatabase, id: MacroCallId) -> ExpandTo {
- db.lookup_intern_macro_call(id).expand_to()
-}
-
fn token_tree_to_syntax_node(
tt: &tt::Subtree,
expand_to: ExpandTo,
-) -> (Parse<SyntaxNode>, mbe::TokenMap) {
+) -> (Parse<SyntaxNode>, ExpansionSpanMap) {
let entry_point = match expand_to {
ExpandTo::Statements => mbe::TopEntryPoint::MacroStmts,
ExpandTo::Items => mbe::TopEntryPoint::MacroItems,
@@ -695,7 +727,7 @@ fn check_tt_count(tt: &tt::Subtree) -> Result<(), ExpandResult<Arc<tt::Subtree>>
if TOKEN_LIMIT.check(count).is_err() {
Err(ExpandResult {
value: Arc::new(tt::Subtree {
- delimiter: tt::Delimiter::UNSPECIFIED,
+ delimiter: tt::Delimiter::DUMMY_INVISIBLE,
token_trees: vec![],
}),
err: Some(ExpandError::other(format!(
@@ -708,3 +740,44 @@ fn check_tt_count(tt: &tt::Subtree) -> Result<(), ExpandResult<Arc<tt::Subtree>>
Ok(())
}
}
+
+fn setup_syntax_context_root(db: &dyn ExpandDatabase) {
+ db.intern_syntax_context(SyntaxContextData::root());
+}
+
+fn dump_syntax_contexts(db: &dyn ExpandDatabase) -> String {
+ let mut s = String::from("Expansions:");
+ let mut entries = InternMacroCallLookupQuery.in_db(db).entries::<Vec<_>>();
+ entries.sort_by_key(|e| e.key);
+ for e in entries {
+ let id = e.key;
+ let expn_data = e.value.as_ref().unwrap();
+ s.push_str(&format!(
+ "\n{:?}: parent: {:?}, call_site_ctxt: {:?}, def_site_ctxt: {:?}, kind: {:?}",
+ id,
+ expn_data.kind.file_id(),
+ expn_data.call_site,
+ SyntaxContextId::ROOT, // FIXME expn_data.def_site,
+ expn_data.kind.descr(),
+ ));
+ }
+
+ s.push_str("\n\nSyntaxContexts:\n");
+ let mut entries = InternSyntaxContextLookupQuery.in_db(db).entries::<Vec<_>>();
+ entries.sort_by_key(|e| e.key);
+ for e in entries {
+ struct SyntaxContextDebug<'a>(
+ &'a dyn ExpandDatabase,
+ SyntaxContextId,
+ &'a SyntaxContextData,
+ );
+
+ impl<'a> std::fmt::Debug for SyntaxContextDebug<'a> {
+ fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
+ self.2.fancy_debug(self.1, self.0, f)
+ }
+ }
+ stdx::format_to!(s, "{:?}\n", SyntaxContextDebug(db, e.key, &e.value.unwrap()));
+ }
+ s
+}