Unnamed repository; edit this file 'description' to name the repository.
Auto merge of #3326 - rust-lang:rustup-2024-02-26, r=RalfJung
Automatic Rustup also fixes https://github.com/rust-lang/miri/issues/3308
bors 2024-02-26
parent 77423ea · parent 893e921 · commit f576d28
-rw-r--r--.github/workflows/ci.yaml5
-rw-r--r--.github/workflows/metrics.yaml9
-rw-r--r--.github/workflows/release.yaml6
-rw-r--r--Cargo.lock1
-rw-r--r--crates/hir-def/src/body/pretty.rs4
-rw-r--r--crates/hir-def/src/data/adt.rs2
-rw-r--r--crates/hir-def/src/import_map.rs2
-rw-r--r--crates/hir-def/src/item_tree.rs6
-rw-r--r--crates/hir-def/src/item_tree/lower.rs28
-rw-r--r--crates/hir-def/src/item_tree/pretty.rs14
-rw-r--r--crates/hir-def/src/nameres/collector.rs2
-rw-r--r--crates/hir-def/src/nameres/tests/macros.rs3
-rw-r--r--crates/hir-ty/src/chalk_db.rs292
-rw-r--r--crates/hir-ty/src/db.rs8
-rw-r--r--crates/hir-ty/src/diagnostics/expr.rs117
-rw-r--r--crates/hir-ty/src/infer.rs3
-rw-r--r--crates/hir-ty/src/infer/expr.rs14
-rw-r--r--crates/hir-ty/src/mir/eval/shim.rs14
-rw-r--r--crates/hir-ty/src/mir/eval/shim/simd.rs1
-rw-r--r--crates/hir-ty/src/mir/lower.rs20
-rw-r--r--crates/hir-ty/src/mir/lower/pattern_matching.rs15
-rw-r--r--crates/hir-ty/src/tests/traits.rs55
-rw-r--r--crates/hir/src/diagnostics.rs33
-rw-r--r--crates/hir/src/lib.rs92
-rw-r--r--crates/hir/src/term_search/tactics.rs6
-rw-r--r--crates/ide-completion/src/item.rs2
-rw-r--r--crates/ide-db/src/apply_change.rs4
-rw-r--r--crates/ide-db/src/defs.rs34
-rw-r--r--crates/ide-db/src/imports/insert_use/tests.rs1
-rw-r--r--crates/ide-db/src/lib.rs2
-rw-r--r--crates/ide-db/src/symbol_index.rs1
-rw-r--r--crates/ide-diagnostics/src/handlers/inactive_code.rs1
-rw-r--r--crates/ide-diagnostics/src/handlers/incorrect_case.rs2
-rw-r--r--crates/ide-diagnostics/src/handlers/missing_fields.rs3
-rw-r--r--crates/ide-diagnostics/src/handlers/missing_match_arms.rs16
-rw-r--r--crates/ide-diagnostics/src/handlers/mutability_errors.rs6
-rw-r--r--crates/ide-diagnostics/src/handlers/non_exhaustive_let.rs47
-rw-r--r--crates/ide-diagnostics/src/handlers/remove_trailing_return.rs2
-rw-r--r--crates/ide-diagnostics/src/handlers/remove_unnecessary_else.rs158
-rw-r--r--crates/ide-diagnostics/src/handlers/type_mismatch.rs2
-rw-r--r--crates/ide-diagnostics/src/handlers/undeclared_label.rs8
-rw-r--r--crates/ide-diagnostics/src/handlers/unresolved_field.rs7
-rw-r--r--crates/ide-diagnostics/src/handlers/unresolved_ident.rs46
-rw-r--r--crates/ide-diagnostics/src/handlers/unresolved_method.rs4
-rw-r--r--crates/ide-diagnostics/src/handlers/useless_braces.rs4
-rw-r--r--crates/ide-diagnostics/src/lib.rs24
-rw-r--r--crates/ide-diagnostics/src/tests.rs7
-rw-r--r--crates/ide/src/hover.rs37
-rw-r--r--crates/ide/src/hover/render.rs36
-rw-r--r--crates/ide/src/hover/tests.rs131
-rw-r--r--crates/ide/src/join_lines.rs1
-rw-r--r--crates/ide/src/rename.rs123
-rw-r--r--crates/ide/src/static_index.rs4
-rw-r--r--crates/load-cargo/src/lib.rs15
-rw-r--r--crates/proc-macro-api/src/process.rs7
-rw-r--r--crates/proc-macro-srv/src/proc_macros.rs6
-rw-r--r--crates/proc-macro-srv/src/server.rs9
-rw-r--r--crates/proc-macro-srv/src/server/rust_analyzer_span.rs47
-rw-r--r--crates/proc-macro-srv/src/server/token_id.rs47
-rw-r--r--crates/proc-macro-srv/src/server/token_stream.rs11
-rw-r--r--crates/proc-macro-srv/src/tests/mod.rs26
-rw-r--r--crates/project-model/src/build_scripts.rs7
-rw-r--r--crates/project-model/src/target_data_layout.rs11
-rw-r--r--crates/project-model/src/workspace.rs96
-rw-r--r--crates/rust-analyzer/src/cargo_target_spec.rs1
-rw-r--r--crates/rust-analyzer/src/cli/flags.rs2
-rw-r--r--crates/rust-analyzer/src/cli/rustc_tests.rs93
-rw-r--r--crates/rust-analyzer/src/cli/scip.rs2
-rw-r--r--crates/rust-analyzer/src/global_state.rs20
-rw-r--r--crates/rust-analyzer/src/handlers/request.rs13
-rw-r--r--crates/rust-analyzer/src/lsp/to_proto.rs23
-rw-r--r--crates/rust-analyzer/src/lsp/utils.rs1
-rw-r--r--crates/rust-analyzer/src/reload.rs51
-rw-r--r--crates/rust-analyzer/tests/slow-tests/support.rs2
-rw-r--r--crates/salsa/Cargo.toml1
-rw-r--r--crates/salsa/salsa-macros/src/database_storage.rs4
-rw-r--r--crates/salsa/salsa-macros/src/query_group.rs4
-rw-r--r--crates/salsa/src/derived.rs28
-rw-r--r--crates/salsa/src/derived/slot.rs124
-rw-r--r--crates/salsa/src/durability.rs4
-rw-r--r--crates/salsa/src/input.rs53
-rw-r--r--crates/salsa/src/interned.rs16
-rw-r--r--crates/salsa/src/lib.rs19
-rw-r--r--crates/salsa/src/lru.rs2
-rw-r--r--crates/salsa/src/plumbing.rs15
-rw-r--r--crates/salsa/src/revision.rs2
-rw-r--r--crates/salsa/src/runtime.rs52
-rw-r--r--crates/salsa/src/runtime/dependency_graph.rs2
-rw-r--r--crates/salsa/src/runtime/local_state.rs7
-rw-r--r--crates/salsa/tests/incremental/memoized_volatile.rs4
-rw-r--r--crates/salsa/tests/on_demand_inputs.rs4
-rw-r--r--crates/salsa/tests/storage_varieties/tests.rs4
-rw-r--r--crates/stdx/src/lib.rs16
-rw-r--r--editors/code/src/commands.ts66
-rw-r--r--editors/code/src/snippets.ts144
-rw-r--r--xtask/src/metrics.rs6
96 files changed, 1828 insertions, 704 deletions
diff --git a/.github/workflows/ci.yaml b/.github/workflows/ci.yaml
index 62fbd57abc..5a8b18e3fe 100644
--- a/.github/workflows/ci.yaml
+++ b/.github/workflows/ci.yaml
@@ -226,6 +226,11 @@ jobs:
- name: download typos
run: curl -LsSf https://github.com/crate-ci/typos/releases/download/$TYPOS_VERSION/typos-$TYPOS_VERSION-x86_64-unknown-linux-musl.tar.gz | tar zxf - -C ${CARGO_HOME:-~/.cargo}/bin
+ - name: Checkout repository
+ uses: actions/checkout@v4
+ with:
+ ref: ${{ github.event.pull_request.head.sha }}
+
- name: check for typos
run: typos
diff --git a/.github/workflows/metrics.yaml b/.github/workflows/metrics.yaml
index be9f504e59..de61b2389a 100644
--- a/.github/workflows/metrics.yaml
+++ b/.github/workflows/metrics.yaml
@@ -67,7 +67,7 @@ jobs:
other_metrics:
strategy:
matrix:
- names: [self, rustc_tests, ripgrep-13.0.0, webrender-2022, diesel-1.4.8, hyper-0.14.18]
+ names: [self, ripgrep-13.0.0, webrender-2022, diesel-1.4.8, hyper-0.14.18]
runs-on: ubuntu-latest
needs: [setup_cargo, build_metrics]
@@ -118,11 +118,6 @@ jobs:
with:
name: self-${{ github.sha }}
- - name: Download rustc_tests metrics
- uses: actions/download-artifact@v3
- with:
- name: rustc_tests-${{ github.sha }}
-
- name: Download ripgrep-13.0.0 metrics
uses: actions/download-artifact@v3
with:
@@ -151,7 +146,7 @@ jobs:
chmod 700 ~/.ssh
git clone --depth 1 [email protected]:rust-analyzer/metrics.git
- jq -s ".[0] * .[1] * .[2] * .[3] * .[4] * .[5] * .[6]" build.json self.json rustc_tests.json ripgrep-13.0.0.json webrender-2022.json diesel-1.4.8.json hyper-0.14.18.json -c >> metrics/metrics.json
+ jq -s ".[0] * .[1] * .[2] * .[3] * .[4] * .[5]" build.json self.json ripgrep-13.0.0.json webrender-2022.json diesel-1.4.8.json hyper-0.14.18.json -c >> metrics/metrics.json
cd metrics
git add .
git -c user.name=Bot -c [email protected] commit --message 📈
diff --git a/.github/workflows/release.yaml b/.github/workflows/release.yaml
index adb1c85051..ac536d0fdd 100644
--- a/.github/workflows/release.yaml
+++ b/.github/workflows/release.yaml
@@ -59,7 +59,7 @@ jobs:
steps:
- name: Checkout repository
- uses: actions/checkout@v4
+ uses: actions/checkout@v3
with:
fetch-depth: ${{ env.FETCH_DEPTH }}
@@ -78,9 +78,9 @@ jobs:
rustup component add rust-src
- name: Install Node.js
- uses: actions/setup-node@v4
+ uses: actions/setup-node@v3
with:
- node-version: 18
+ node-version: 16
- name: Update apt repositories
if: matrix.target == 'aarch64-unknown-linux-gnu' || matrix.target == 'arm-unknown-linux-gnueabihf'
diff --git a/Cargo.lock b/Cargo.lock
index 7b29d7bb79..3c87291dba 100644
--- a/Cargo.lock
+++ b/Cargo.lock
@@ -1709,6 +1709,7 @@ dependencies = [
"dissimilar",
"expect-test",
"indexmap",
+ "itertools",
"linked-hash-map",
"lock_api",
"oorandom",
diff --git a/crates/hir-def/src/body/pretty.rs b/crates/hir-def/src/body/pretty.rs
index 8229b1ccf3..cd14f7b855 100644
--- a/crates/hir-def/src/body/pretty.rs
+++ b/crates/hir-def/src/body/pretty.rs
@@ -6,8 +6,8 @@ use itertools::Itertools;
use crate::{
hir::{
- Array, BindingAnnotation, CaptureBy, ClosureKind, Literal, LiteralOrConst,
- Movability, Statement,
+ Array, BindingAnnotation, CaptureBy, ClosureKind, Literal, LiteralOrConst, Movability,
+ Statement,
},
pretty::{print_generic_args, print_path, print_type_ref},
type_ref::TypeRef,
diff --git a/crates/hir-def/src/data/adt.rs b/crates/hir-def/src/data/adt.rs
index 540f643ae7..f07b125766 100644
--- a/crates/hir-def/src/data/adt.rs
+++ b/crates/hir-def/src/data/adt.rs
@@ -40,7 +40,7 @@ pub struct StructData {
}
bitflags! {
- #[derive(Debug, Clone, PartialEq, Eq)]
+ #[derive(Debug, Copy, Clone, PartialEq, Eq)]
pub struct StructFlags: u8 {
const NO_FLAGS = 0;
/// Indicates whether the struct is `PhantomData`.
diff --git a/crates/hir-def/src/import_map.rs b/crates/hir-def/src/import_map.rs
index 38cfcf0f28..faa1eed15a 100644
--- a/crates/hir-def/src/import_map.rs
+++ b/crates/hir-def/src/import_map.rs
@@ -477,7 +477,7 @@ mod tests {
use expect_test::{expect, Expect};
use test_fixture::WithFixture;
- use crate::{db::DefDatabase, test_db::TestDB, ItemContainerId, Lookup};
+ use crate::{test_db::TestDB, ItemContainerId, Lookup};
use super::*;
diff --git a/crates/hir-def/src/item_tree.rs b/crates/hir-def/src/item_tree.rs
index be16a5e31a..bb36950f95 100644
--- a/crates/hir-def/src/item_tree.rs
+++ b/crates/hir-def/src/item_tree.rs
@@ -44,13 +44,13 @@ use std::{
ops::{Index, Range},
};
-use ast::{AstNode, HasName, StructKind};
+use ast::{AstNode, StructKind};
use base_db::CrateId;
use either::Either;
use hir_expand::{
ast_id_map::{AstIdNode, FileAstId},
attrs::RawAttrs,
- name::{name, AsName, Name},
+ name::Name,
ExpandTo, HirFileId, InFile,
};
use intern::Interned;
@@ -67,7 +67,7 @@ use crate::{
attr::Attrs,
db::DefDatabase,
generics::{GenericParams, LifetimeParamData, TypeOrConstParamData},
- path::{path, AssociatedTypeBinding, GenericArgs, ImportAlias, ModPath, Path, PathKind},
+ path::{GenericArgs, ImportAlias, ModPath, Path, PathKind},
type_ref::{Mutability, TraitRef, TypeBound, TypeRef},
visibility::{RawVisibility, VisibilityExplicitness},
BlockId, Lookup,
diff --git a/crates/hir-def/src/item_tree/lower.rs b/crates/hir-def/src/item_tree/lower.rs
index b51cb5de0f..37fdece876 100644
--- a/crates/hir-def/src/item_tree/lower.rs
+++ b/crates/hir-def/src/item_tree/lower.rs
@@ -2,17 +2,33 @@
use std::collections::hash_map::Entry;
-use hir_expand::{ast_id_map::AstIdMap, span_map::SpanMapRef};
-use syntax::ast::{HasModuleItem, HasTypeBounds, IsString};
+use hir_expand::{
+ ast_id_map::AstIdMap, mod_path::path, name, name::AsName, span_map::SpanMapRef, HirFileId,
+};
+use la_arena::Arena;
+use syntax::{
+ ast::{self, HasModuleItem, HasName, HasTypeBounds, IsString},
+ AstNode,
+};
+use triomphe::Arc;
use crate::{
- generics::{GenericParamsCollector, TypeParamData, TypeParamProvenance},
- type_ref::{LifetimeRef, TraitBoundModifier},
+ db::DefDatabase,
+ generics::{GenericParams, GenericParamsCollector, TypeParamData, TypeParamProvenance},
+ item_tree::{
+ AssocItem, AttrOwner, Const, Either, Enum, ExternBlock, ExternCrate, Field, FieldAstId,
+ Fields, FileItemTreeId, FnFlags, Function, GenericArgs, Idx, IdxRange, Impl, ImportAlias,
+ Interned, ItemTree, ItemTreeData, ItemTreeNode, Macro2, MacroCall, MacroRules, Mod,
+ ModItem, ModKind, ModPath, Mutability, Name, Param, ParamAstId, Path, Range, RawAttrs,
+ RawIdx, RawVisibilityId, Static, Struct, StructKind, Trait, TraitAlias, TypeAlias, Union,
+ Use, UseTree, UseTreeKind, Variant,
+ },
+ path::AssociatedTypeBinding,
+ type_ref::{LifetimeRef, TraitBoundModifier, TraitRef, TypeBound, TypeRef},
+ visibility::RawVisibility,
LocalLifetimeParamId, LocalTypeOrConstParamId,
};
-use super::*;
-
fn id<N: ItemTreeNode>(index: Idx<N>) -> FileItemTreeId<N> {
FileItemTreeId(index)
}
diff --git a/crates/hir-def/src/item_tree/pretty.rs b/crates/hir-def/src/item_tree/pretty.rs
index dae876f7ec..87c90a4c6a 100644
--- a/crates/hir-def/src/item_tree/pretty.rs
+++ b/crates/hir-def/src/item_tree/pretty.rs
@@ -1,16 +1,22 @@
//! `ItemTree` debug printer.
-use std::fmt::Write;
+use std::fmt::{self, Write};
use span::ErasedFileAstId;
use crate::{
- generics::{WherePredicate, WherePredicateTypeTarget},
+ generics::{TypeOrConstParamData, WherePredicate, WherePredicateTypeTarget},
+ item_tree::{
+ AttrOwner, Const, DefDatabase, Enum, ExternBlock, ExternCrate, Field, FieldAstId, Fields,
+ FileItemTreeId, FnFlags, Function, GenericParams, Impl, Interned, ItemTree, Macro2,
+ MacroCall, MacroRules, Mod, ModItem, ModKind, Param, ParamAstId, Path, RawAttrs,
+ RawVisibilityId, Static, Struct, Trait, TraitAlias, TypeAlias, TypeBound, TypeRef, Union,
+ Use, UseTree, UseTreeKind, Variant,
+ },
pretty::{print_path, print_type_bounds, print_type_ref},
+ visibility::RawVisibility,
};
-use super::*;
-
pub(super) fn print_item_tree(db: &dyn DefDatabase, tree: &ItemTree) -> String {
let mut p = Printer { db, tree, buf: String::new(), indent_level: 0, needs_indent: true };
diff --git a/crates/hir-def/src/nameres/collector.rs b/crates/hir-def/src/nameres/collector.rs
index 88838f58fe..3282540650 100644
--- a/crates/hir-def/src/nameres/collector.rs
+++ b/crates/hir-def/src/nameres/collector.rs
@@ -2446,7 +2446,7 @@ mod tests {
use base_db::SourceDatabase;
use test_fixture::WithFixture;
- use crate::{db::DefDatabase, test_db::TestDB};
+ use crate::test_db::TestDB;
use super::*;
diff --git a/crates/hir-def/src/nameres/tests/macros.rs b/crates/hir-def/src/nameres/tests/macros.rs
index bf89ea711a..d278b75e81 100644
--- a/crates/hir-def/src/nameres/tests/macros.rs
+++ b/crates/hir-def/src/nameres/tests/macros.rs
@@ -1,10 +1,7 @@
use expect_test::expect;
-use test_fixture::WithFixture;
use itertools::Itertools;
-use crate::nameres::tests::check;
-
use super::*;
#[test]
diff --git a/crates/hir-ty/src/chalk_db.rs b/crates/hir-ty/src/chalk_db.rs
index bd243518fc..e678a2fee1 100644
--- a/crates/hir-ty/src/chalk_db.rs
+++ b/crates/hir-ty/src/chalk_db.rs
@@ -1,7 +1,7 @@
//! The implementation of `RustIrDatabase` for Chalk, which provides information
//! about the code that Chalk needs.
use core::ops;
-use std::{iter, sync::Arc};
+use std::{iter, ops::ControlFlow, sync::Arc};
use tracing::debug;
@@ -10,9 +10,10 @@ use chalk_solve::rust_ir::{self, OpaqueTyDatumBound, WellKnownTrait};
use base_db::CrateId;
use hir_def::{
+ data::adt::StructFlags,
hir::Movability,
lang_item::{LangItem, LangItemTarget},
- AssocItemId, BlockId, GenericDefId, HasModule, ItemContainerId, Lookup, TypeAliasId,
+ AssocItemId, BlockId, GenericDefId, HasModule, ItemContainerId, Lookup, TypeAliasId, VariantId,
};
use hir_expand::name::name;
@@ -33,7 +34,7 @@ use crate::{
pub(crate) type AssociatedTyDatum = chalk_solve::rust_ir::AssociatedTyDatum<Interner>;
pub(crate) type TraitDatum = chalk_solve::rust_ir::TraitDatum<Interner>;
-pub(crate) type StructDatum = chalk_solve::rust_ir::AdtDatum<Interner>;
+pub(crate) type AdtDatum = chalk_solve::rust_ir::AdtDatum<Interner>;
pub(crate) type ImplDatum = chalk_solve::rust_ir::ImplDatum<Interner>;
pub(crate) type OpaqueTyDatum = chalk_solve::rust_ir::OpaqueTyDatum<Interner>;
@@ -53,8 +54,8 @@ impl chalk_solve::RustIrDatabase<Interner> for ChalkContext<'_> {
fn trait_datum(&self, trait_id: TraitId) -> Arc<TraitDatum> {
self.db.trait_datum(self.krate, trait_id)
}
- fn adt_datum(&self, struct_id: AdtId) -> Arc<StructDatum> {
- self.db.struct_datum(self.krate, struct_id)
+ fn adt_datum(&self, struct_id: AdtId) -> Arc<AdtDatum> {
+ self.db.adt_datum(self.krate, struct_id)
}
fn adt_repr(&self, _struct_id: AdtId) -> Arc<rust_ir::AdtRepr<Interner>> {
// FIXME: keep track of these
@@ -136,81 +137,92 @@ impl chalk_solve::RustIrDatabase<Interner> for ChalkContext<'_> {
_ => self_ty_fp.as_ref().map(std::slice::from_ref).unwrap_or(&[]),
};
- let trait_module = trait_.module(self.db.upcast());
- let type_module = match self_ty_fp {
- Some(TyFingerprint::Adt(adt_id)) => Some(adt_id.module(self.db.upcast())),
- Some(TyFingerprint::ForeignType(type_id)) => {
- Some(from_foreign_def_id(type_id).module(self.db.upcast()))
- }
- Some(TyFingerprint::Dyn(trait_id)) => Some(trait_id.module(self.db.upcast())),
- _ => None,
- };
-
- let mut def_blocks =
- [trait_module.containing_block(), type_module.and_then(|it| it.containing_block())];
-
- // Note: Since we're using impls_for_trait, only impls where the trait
- // can be resolved should ever reach Chalk. impl_datum relies on that
- // and will panic if the trait can't be resolved.
- let in_deps = self.db.trait_impls_in_deps(self.krate);
- let in_self = self.db.trait_impls_in_crate(self.krate);
-
- let block_impls = iter::successors(self.block, |&block_id| {
- cov_mark::hit!(block_local_impls);
- self.db.block_def_map(block_id).parent().and_then(|module| module.containing_block())
- })
- .inspect(|&block_id| {
- // make sure we don't search the same block twice
- def_blocks.iter_mut().for_each(|block| {
- if *block == Some(block_id) {
- *block = None;
- }
- });
- })
- .filter_map(|block_id| self.db.trait_impls_in_block(block_id));
-
let id_to_chalk = |id: hir_def::ImplId| id.to_chalk(self.db);
+
let mut result = vec![];
- match fps {
- [] => {
- debug!("Unrestricted search for {:?} impls...", trait_);
- let mut f = |impls: &TraitImpls| {
- result.extend(impls.for_trait(trait_).map(id_to_chalk));
- };
- f(&in_self);
- in_deps.iter().map(ops::Deref::deref).for_each(&mut f);
- block_impls.for_each(|it| f(&it));
- def_blocks
- .into_iter()
- .flatten()
- .filter_map(|it| self.db.trait_impls_in_block(it))
- .for_each(|it| f(&it));
- }
- fps => {
- let mut f =
- |impls: &TraitImpls| {
- result.extend(fps.iter().flat_map(|fp| {
- impls.for_trait_and_self_ty(trait_, *fp).map(id_to_chalk)
- }));
- };
- f(&in_self);
- in_deps.iter().map(ops::Deref::deref).for_each(&mut f);
- block_impls.for_each(|it| f(&it));
- def_blocks
- .into_iter()
- .flatten()
- .filter_map(|it| self.db.trait_impls_in_block(it))
- .for_each(|it| f(&it));
- }
- }
+ if fps.is_empty() {
+ debug!("Unrestricted search for {:?} impls...", trait_);
+ self.for_trait_impls(trait_, self_ty_fp, |impls| {
+ result.extend(impls.for_trait(trait_).map(id_to_chalk));
+ ControlFlow::Continue(())
+ })
+ } else {
+ self.for_trait_impls(trait_, self_ty_fp, |impls| {
+ result.extend(
+ fps.iter().flat_map(move |fp| {
+ impls.for_trait_and_self_ty(trait_, *fp).map(id_to_chalk)
+ }),
+ );
+ ControlFlow::Continue(())
+ })
+ };
debug!("impls_for_trait returned {} impls", result.len());
result
}
+
fn impl_provided_for(&self, auto_trait_id: TraitId, kind: &chalk_ir::TyKind<Interner>) -> bool {
debug!("impl_provided_for {:?}, {:?}", auto_trait_id, kind);
- false // FIXME
+
+ let trait_id = from_chalk_trait_id(auto_trait_id);
+ let self_ty = kind.clone().intern(Interner);
+ // We cannot filter impls by `TyFingerprint` for the following types:
+ let self_ty_fp = match kind {
+ // because we need to find any impl whose Self type is a ref with the same mutability
+ // (we don't care about the inner type).
+ TyKind::Ref(..) => None,
+ // because we need to find any impl whose Self type is a tuple with the same arity.
+ TyKind::Tuple(..) => None,
+ _ => TyFingerprint::for_trait_impl(&self_ty),
+ };
+
+ let check_kind = |impl_id| {
+ let impl_self_ty = self.db.impl_self_ty(impl_id);
+ // NOTE(skip_binders): it's safe to skip binders here as we don't check substitutions.
+ let impl_self_kind = impl_self_ty.skip_binders().kind(Interner);
+
+ match (kind, impl_self_kind) {
+ (TyKind::Adt(id_a, _), TyKind::Adt(id_b, _)) => id_a == id_b,
+ (TyKind::AssociatedType(id_a, _), TyKind::AssociatedType(id_b, _)) => id_a == id_b,
+ (TyKind::Scalar(scalar_a), TyKind::Scalar(scalar_b)) => scalar_a == scalar_b,
+ (TyKind::Error, TyKind::Error)
+ | (TyKind::Str, TyKind::Str)
+ | (TyKind::Slice(_), TyKind::Slice(_))
+ | (TyKind::Never, TyKind::Never)
+ | (TyKind::Array(_, _), TyKind::Array(_, _)) => true,
+ (TyKind::Tuple(arity_a, _), TyKind::Tuple(arity_b, _)) => arity_a == arity_b,
+ (TyKind::OpaqueType(id_a, _), TyKind::OpaqueType(id_b, _)) => id_a == id_b,
+ (TyKind::FnDef(id_a, _), TyKind::FnDef(id_b, _)) => id_a == id_b,
+ (TyKind::Ref(id_a, _, _), TyKind::Ref(id_b, _, _))
+ | (TyKind::Raw(id_a, _), TyKind::Raw(id_b, _)) => id_a == id_b,
+ (TyKind::Closure(id_a, _), TyKind::Closure(id_b, _)) => id_a == id_b,
+ (TyKind::Coroutine(id_a, _), TyKind::Coroutine(id_b, _))
+ | (TyKind::CoroutineWitness(id_a, _), TyKind::CoroutineWitness(id_b, _)) => {
+ id_a == id_b
+ }
+ (TyKind::Foreign(id_a), TyKind::Foreign(id_b)) => id_a == id_b,
+ (_, _) => false,
+ }
+ };
+
+ if let Some(fp) = self_ty_fp {
+ self.for_trait_impls(trait_id, self_ty_fp, |impls| {
+ match impls.for_trait_and_self_ty(trait_id, fp).any(check_kind) {
+ true => ControlFlow::Break(()),
+ false => ControlFlow::Continue(()),
+ }
+ })
+ } else {
+ self.for_trait_impls(trait_id, self_ty_fp, |impls| {
+ match impls.for_trait(trait_id).any(check_kind) {
+ true => ControlFlow::Break(()),
+ false => ControlFlow::Continue(()),
+ }
+ })
+ }
+ .is_break()
}
+
fn associated_ty_value(&self, id: AssociatedTyValueId) -> Arc<AssociatedTyValue> {
self.db.associated_ty_value(self.krate, id)
}
@@ -489,6 +501,59 @@ impl chalk_solve::RustIrDatabase<Interner> for ChalkContext<'_> {
}
}
+impl<'a> ChalkContext<'a> {
+ fn for_trait_impls(
+ &self,
+ trait_id: hir_def::TraitId,
+ self_ty_fp: Option<TyFingerprint>,
+ mut f: impl FnMut(&TraitImpls) -> ControlFlow<()>,
+ ) -> ControlFlow<()> {
+ // Note: Since we're using `impls_for_trait` and `impl_provided_for`,
+ // only impls where the trait can be resolved should ever reach Chalk.
+ // `impl_datum` relies on that and will panic if the trait can't be resolved.
+ let in_deps = self.db.trait_impls_in_deps(self.krate);
+ let in_self = self.db.trait_impls_in_crate(self.krate);
+ let trait_module = trait_id.module(self.db.upcast());
+ let type_module = match self_ty_fp {
+ Some(TyFingerprint::Adt(adt_id)) => Some(adt_id.module(self.db.upcast())),
+ Some(TyFingerprint::ForeignType(type_id)) => {
+ Some(from_foreign_def_id(type_id).module(self.db.upcast()))
+ }
+ Some(TyFingerprint::Dyn(trait_id)) => Some(trait_id.module(self.db.upcast())),
+ _ => None,
+ };
+
+ let mut def_blocks =
+ [trait_module.containing_block(), type_module.and_then(|it| it.containing_block())];
+
+ let block_impls = iter::successors(self.block, |&block_id| {
+ cov_mark::hit!(block_local_impls);
+ self.db.block_def_map(block_id).parent().and_then(|module| module.containing_block())
+ })
+ .inspect(|&block_id| {
+ // make sure we don't search the same block twice
+ def_blocks.iter_mut().for_each(|block| {
+ if *block == Some(block_id) {
+ *block = None;
+ }
+ });
+ })
+ .filter_map(|block_id| self.db.trait_impls_in_block(block_id));
+ f(&in_self)?;
+ for it in in_deps.iter().map(ops::Deref::deref) {
+ f(it)?;
+ }
+ for it in block_impls {
+ f(&it)?;
+ }
+ for it in def_blocks.into_iter().flatten().filter_map(|it| self.db.trait_impls_in_block(it))
+ {
+ f(&it)?;
+ }
+ ControlFlow::Continue(())
+ }
+}
+
impl chalk_ir::UnificationDatabase<Interner> for &dyn HirDatabase {
fn fn_def_variance(
&self,
@@ -590,7 +655,7 @@ pub(crate) fn trait_datum_query(
coinductive: false, // only relevant for Chalk testing
// FIXME: set these flags correctly
marker: false,
- fundamental: false,
+ fundamental: trait_data.fundamental,
};
let where_clauses = convert_where_clauses(db, trait_.into(), &bound_vars);
let associated_ty_ids = trait_data.associated_types().map(to_assoc_type_id).collect();
@@ -649,35 +714,74 @@ fn lang_item_from_well_known_trait(trait_: WellKnownTrait) -> LangItem {
}
}
-pub(crate) fn struct_datum_query(
+pub(crate) fn adt_datum_query(
db: &dyn HirDatabase,
krate: CrateId,
- struct_id: AdtId,
-) -> Arc<StructDatum> {
- debug!("struct_datum {:?}", struct_id);
- let chalk_ir::AdtId(adt_id) = struct_id;
+ chalk_ir::AdtId(adt_id): AdtId,
+) -> Arc<AdtDatum> {
+ debug!("adt_datum {:?}", adt_id);
let generic_params = generics(db.upcast(), adt_id.into());
- let upstream = adt_id.module(db.upcast()).krate() != krate;
- let where_clauses = {
- let generic_params = generics(db.upcast(), adt_id.into());
- let bound_vars = generic_params.bound_vars_subst(db, DebruijnIndex::INNERMOST);
- convert_where_clauses(db, adt_id.into(), &bound_vars)
+ let bound_vars_subst = generic_params.bound_vars_subst(db, DebruijnIndex::INNERMOST);
+ let where_clauses = convert_where_clauses(db, adt_id.into(), &bound_vars_subst);
+
+ let (fundamental, phantom_data) = match adt_id {
+ hir_def::AdtId::StructId(s) => {
+ let flags = db.struct_data(s).flags;
+ (
+ flags.contains(StructFlags::IS_FUNDAMENTAL),
+ flags.contains(StructFlags::IS_PHANTOM_DATA),
+ )
+ }
+ // FIXME set fundamental flags correctly
+ hir_def::AdtId::UnionId(_) => (false, false),
+ hir_def::AdtId::EnumId(_) => (false, false),
};
let flags = rust_ir::AdtFlags {
- upstream,
- // FIXME set fundamental and phantom_data flags correctly
- fundamental: false,
- phantom_data: false,
+ upstream: adt_id.module(db.upcast()).krate() != krate,
+ fundamental,
+ phantom_data,
};
- // FIXME provide enum variants properly (for auto traits)
- let variant = rust_ir::AdtVariantDatum {
- fields: Vec::new(), // FIXME add fields (only relevant for auto traits),
+
+ // this slows down rust-analyzer by quite a bit unfortunately, so enabling this is currently not worth it
+ let _variant_id_to_fields = |id: VariantId| {
+ let variant_data = &id.variant_data(db.upcast());
+ let fields = if variant_data.fields().is_empty() {
+ vec![]
+ } else {
+ let field_types = db.field_types(id);
+ variant_data
+ .fields()
+ .iter()
+ .map(|(idx, _)| field_types[idx].clone().substitute(Interner, &bound_vars_subst))
+ .filter(|it| !it.contains_unknown())
+ .collect()
+ };
+ rust_ir::AdtVariantDatum { fields }
};
- let struct_datum_bound = rust_ir::AdtDatumBound { variants: vec![variant], where_clauses };
- let struct_datum = StructDatum {
- // FIXME set ADT kind
- kind: rust_ir::AdtKind::Struct,
- id: struct_id,
+ let variant_id_to_fields = |_: VariantId| rust_ir::AdtVariantDatum { fields: vec![] };
+
+ let (kind, variants) = match adt_id {
+ hir_def::AdtId::StructId(id) => {
+ (rust_ir::AdtKind::Struct, vec![variant_id_to_fields(id.into())])
+ }
+ hir_def::AdtId::EnumId(id) => {
+ let variants = db
+ .enum_data(id)
+ .variants
+ .iter()
+ .map(|&(variant_id, _)| variant_id_to_fields(variant_id.into()))
+ .collect();
+ (rust_ir::AdtKind::Enum, variants)
+ }
+ hir_def::AdtId::UnionId(id) => {
+ (rust_ir::AdtKind::Union, vec![variant_id_to_fields(id.into())])
+ }
+ };
+
+ let struct_datum_bound = rust_ir::AdtDatumBound { variants, where_clauses };
+ let struct_datum = AdtDatum {
+ kind,
+ id: chalk_ir::AdtId(adt_id),
binders: make_binders(db, &generic_params, struct_datum_bound),
flags,
};
diff --git a/crates/hir-ty/src/db.rs b/crates/hir-ty/src/db.rs
index fbd366864a..f9e8cff553 100644
--- a/crates/hir-ty/src/db.rs
+++ b/crates/hir-ty/src/db.rs
@@ -90,7 +90,7 @@ pub trait HirDatabase: DefDatabase + Upcast<dyn DefDatabase> {
#[salsa::cycle(crate::lower::ty_recover)]
fn ty(&self, def: TyDefId) -> Binders<Ty>;
- /// Returns the type of the value of the given constant, or `None` if the the `ValueTyDefId` is
+ /// Returns the type of the value of the given constant, or `None` if the `ValueTyDefId` is
/// a `StructId` or `EnumVariantId` with a record constructor.
#[salsa::invoke(crate::lower::value_ty_query)]
fn value_ty(&self, def: ValueTyDefId) -> Option<Binders<Ty>>;
@@ -220,12 +220,12 @@ pub trait HirDatabase: DefDatabase + Upcast<dyn DefDatabase> {
trait_id: chalk_db::TraitId,
) -> sync::Arc<chalk_db::TraitDatum>;
- #[salsa::invoke(chalk_db::struct_datum_query)]
- fn struct_datum(
+ #[salsa::invoke(chalk_db::adt_datum_query)]
+ fn adt_datum(
&self,
krate: CrateId,
struct_id: chalk_db::AdtId,
- ) -> sync::Arc<chalk_db::StructDatum>;
+ ) -> sync::Arc<chalk_db::AdtDatum>;
#[salsa::invoke(chalk_db::impl_datum_query)]
fn impl_datum(
diff --git a/crates/hir-ty/src/diagnostics/expr.rs b/crates/hir-ty/src/diagnostics/expr.rs
index c4329a7b82..6c8a187516 100644
--- a/crates/hir-ty/src/diagnostics/expr.rs
+++ b/crates/hir-ty/src/diagnostics/expr.rs
@@ -12,6 +12,8 @@ use hir_expand::name;
use itertools::Itertools;
use rustc_hash::FxHashSet;
use rustc_pattern_analysis::usefulness::{compute_match_usefulness, ValidityConstraint};
+use syntax::{ast, AstNode};
+use tracing::debug;
use triomphe::Arc;
use typed_arena::Arena;
@@ -44,6 +46,10 @@ pub enum BodyValidationDiagnostic {
match_expr: ExprId,
uncovered_patterns: String,
},
+ NonExhaustiveLet {
+ pat: PatId,
+ uncovered_patterns: String,
+ },
RemoveTrailingReturn {
return_expr: ExprId,
},
@@ -57,7 +63,8 @@ impl BodyValidationDiagnostic {
let _p =
tracing::span!(tracing::Level::INFO, "BodyValidationDiagnostic::collect").entered();
let infer = db.infer(owner);
- let mut validator = ExprValidator::new(owner, infer);
+ let body = db.body(owner);
+ let mut validator = ExprValidator { owner, body, infer, diagnostics: Vec::new() };
validator.validate_body(db);
validator.diagnostics
}
@@ -65,18 +72,16 @@ impl BodyValidationDiagnostic {
struct ExprValidator {
owner: DefWithBodyId,
+ body: Arc<Body>,
infer: Arc<InferenceResult>,
- pub(super) diagnostics: Vec<BodyValidationDiagnostic>,
+ diagnostics: Vec<BodyValidationDiagnostic>,
}
impl ExprValidator {
- fn new(owner: DefWithBodyId, infer: Arc<InferenceResult>) -> ExprValidator {
- ExprValidator { owner, infer, diagnostics: Vec::new() }
- }
-
fn validate_body(&mut self, db: &dyn HirDatabase) {
- let body = db.body(self.owner);
let mut filter_map_next_checker = None;
+ // we'll pass &mut self while iterating over body.exprs, so they need to be disjoint
+ let body = Arc::clone(&self.body);
if matches!(self.owner, DefWithBodyId::FunctionId(_)) {
self.check_for_trailing_return(body.body_expr, &body);
@@ -104,7 +109,10 @@ impl ExprValidator {
self.check_for_trailing_return(*body_expr, &body);
}
Expr::If { .. } => {
- self.check_for_unnecessary_else(id, expr, &body);
+ self.check_for_unnecessary_else(id, expr, db);
+ }
+ Expr::Block { .. } => {
+ self.validate_block(db, expr);
}
_ => {}
}
@@ -162,8 +170,6 @@ impl ExprValidator {
arms: &[MatchArm],
db: &dyn HirDatabase,
) {
- let body = db.body(self.owner);
-
let scrut_ty = &self.infer[scrutinee_expr];
if scrut_ty.is_unknown() {
return;
@@ -191,12 +197,12 @@ impl ExprValidator {
.as_reference()
.map(|(match_expr_ty, ..)| match_expr_ty == pat_ty)
.unwrap_or(false))
- && types_of_subpatterns_do_match(arm.pat, &body, &self.infer)
+ && types_of_subpatterns_do_match(arm.pat, &self.body, &self.infer)
{
// If we had a NotUsefulMatchArm diagnostic, we could
// check the usefulness of each pattern as we added it
// to the matrix here.
- let pat = self.lower_pattern(&cx, arm.pat, db, &body, &mut has_lowering_errors);
+ let pat = self.lower_pattern(&cx, arm.pat, db, &mut has_lowering_errors);
let m_arm = pat_analysis::MatchArm {
pat: pattern_arena.alloc(pat),
has_guard: arm.guard.is_some(),
@@ -234,20 +240,63 @@ impl ExprValidator {
if !witnesses.is_empty() {
self.diagnostics.push(BodyValidationDiagnostic::MissingMatchArms {
match_expr,
- uncovered_patterns: missing_match_arms(&cx, scrut_ty, witnesses, arms),
+ uncovered_patterns: missing_match_arms(&cx, scrut_ty, witnesses, m_arms.is_empty()),
});
}
}
+ fn validate_block(&mut self, db: &dyn HirDatabase, expr: &Expr) {
+ let Expr::Block { statements, .. } = expr else { return };
+ let pattern_arena = Arena::new();
+ let cx = MatchCheckCtx::new(self.owner.module(db.upcast()), self.owner, db);
+ for stmt in &**statements {
+ let &Statement::Let { pat, initializer, else_branch: None, .. } = stmt else {
+ continue;
+ };
+ let Some(initializer) = initializer else { continue };
+ let ty = &self.infer[initializer];
+
+ let mut have_errors = false;
+ let deconstructed_pat = self.lower_pattern(&cx, pat, db, &mut have_errors);
+ let match_arm = rustc_pattern_analysis::MatchArm {
+ pat: pattern_arena.alloc(deconstructed_pat),
+ has_guard: false,
+ arm_data: (),
+ };
+ if have_errors {
+ continue;
+ }
+
+ let report = match compute_match_usefulness(
+ &cx,
+ &[match_arm],
+ ty.clone(),
+ ValidityConstraint::ValidOnly,
+ ) {
+ Ok(v) => v,
+ Err(e) => {
+ debug!(?e, "match usefulness error");
+ continue;
+ }
+ };
+ let witnesses = report.non_exhaustiveness_witnesses;
+ if !witnesses.is_empty() {
+ self.diagnostics.push(BodyValidationDiagnostic::NonExhaustiveLet {
+ pat,
+ uncovered_patterns: missing_match_arms(&cx, ty, witnesses, false),
+ });
+ }
+ }
+ }
+
fn lower_pattern<'p>(
&self,
cx: &MatchCheckCtx<'p>,
pat: PatId,
db: &dyn HirDatabase,
- body: &Body,
have_errors: &mut bool,
) -> DeconstructedPat<'p> {
- let mut patcx = match_check::PatCtxt::new(db, &self.infer, body);
+ let mut patcx = match_check::PatCtxt::new(db, &self.infer, &self.body);
let pattern = patcx.lower_pattern(pat);
let pattern = cx.lower_pat(&pattern);
if !patcx.errors.is_empty() {
@@ -288,12 +337,12 @@ impl ExprValidator {
}
}
- fn check_for_unnecessary_else(&mut self, id: ExprId, expr: &Expr, body: &Body) {
+ fn check_for_unnecessary_else(&mut self, id: ExprId, expr: &Expr, db: &dyn HirDatabase) {
if let Expr::If { condition: _, then_branch, else_branch } = expr {
if else_branch.is_none() {
return;
}
- if let Expr::Block { statements, tail, .. } = &body.exprs[*then_branch] {
+ if let Expr::Block { statements, tail, .. } = &self.body.exprs[*then_branch] {
let last_then_expr = tail.or_else(|| match statements.last()? {
Statement::Expr { expr, .. } => Some(*expr),
_ => None,
@@ -301,6 +350,36 @@ impl ExprValidator {
if let Some(last_then_expr) = last_then_expr {
let last_then_expr_ty = &self.infer[last_then_expr];
if last_then_expr_ty.is_never() {
+ // Only look at sources if the then branch diverges and we have an else branch.
+ let (_, source_map) = db.body_with_source_map(self.owner);
+ let Ok(source_ptr) = source_map.expr_syntax(id) else {
+ return;
+ };
+ let root = source_ptr.file_syntax(db.upcast());
+ let ast::Expr::IfExpr(if_expr) = source_ptr.value.to_node(&root) else {
+ return;
+ };
+ let mut top_if_expr = if_expr;
+ loop {
+ let parent = top_if_expr.syntax().parent();
+ let has_parent_expr_stmt_or_stmt_list =
+ parent.as_ref().map_or(false, |node| {
+ ast::ExprStmt::can_cast(node.kind())
+ | ast::StmtList::can_cast(node.kind())
+ });
+ if has_parent_expr_stmt_or_stmt_list {
+ // Only emit diagnostic if parent or direct ancestor is either
+ // an expr stmt or a stmt list.
+ break;
+ }
+ let Some(parent_if_expr) = parent.and_then(ast::IfExpr::cast) else {
+ // Bail if parent is neither an if expr, an expr stmt nor a stmt list.
+ return;
+ };
+ // Check parent if expr.
+ top_if_expr = parent_if_expr;
+ }
+
self.diagnostics
.push(BodyValidationDiagnostic::RemoveUnnecessaryElse { if_expr: id })
}
@@ -448,7 +527,7 @@ fn missing_match_arms<'p>(
cx: &MatchCheckCtx<'p>,
scrut_ty: &Ty,
witnesses: Vec<WitnessPat<'p>>,
- arms: &[MatchArm],
+ arms_is_empty: bool,
) -> String {
struct DisplayWitness<'a, 'p>(&'a WitnessPat<'p>, &'a MatchCheckCtx<'p>);
impl fmt::Display for DisplayWitness<'_, '_> {
@@ -463,7 +542,7 @@ fn missing_match_arms<'p>(
Some((AdtId::EnumId(e), _)) => !cx.db.enum_data(e).variants.is_empty(),
_ => false,
};
- if arms.is_empty() && !non_empty_enum {
+ if arms_is_empty && !non_empty_enum {
format!("type `{}` is non-empty", scrut_ty.display(cx.db))
} else {
let pat_display = |witness| DisplayWitness(witness, cx);
diff --git a/crates/hir-ty/src/infer.rs b/crates/hir-ty/src/infer.rs
index 1977f00517..9cea414e1a 100644
--- a/crates/hir-ty/src/infer.rs
+++ b/crates/hir-ty/src/infer.rs
@@ -221,6 +221,9 @@ pub enum InferenceDiagnostic {
UnresolvedAssocItem {
id: ExprOrPatId,
},
+ UnresolvedIdent {
+ expr: ExprId,
+ },
// FIXME: This should be emitted in body lowering
BreakOutsideOfLoop {
expr: ExprId,
diff --git a/crates/hir-ty/src/infer/expr.rs b/crates/hir-ty/src/infer/expr.rs
index 428ed6748c..c377a51e7d 100644
--- a/crates/hir-ty/src/infer/expr.rs
+++ b/crates/hir-ty/src/infer/expr.rs
@@ -13,7 +13,7 @@ use hir_def::{
ArithOp, Array, BinaryOp, ClosureKind, Expr, ExprId, LabelId, Literal, Statement, UnaryOp,
},
lang_item::{LangItem, LangItemTarget},
- path::{GenericArg, GenericArgs},
+ path::{GenericArg, GenericArgs, Path},
BlockId, ConstParamId, FieldId, ItemContainerId, Lookup, TupleFieldId, TupleId,
};
use hir_expand::name::{name, Name};
@@ -439,7 +439,17 @@ impl InferenceContext<'_> {
}
Expr::Path(p) => {
let g = self.resolver.update_to_inner_scope(self.db.upcast(), self.owner, tgt_expr);
- let ty = self.infer_path(p, tgt_expr.into()).unwrap_or_else(|| self.err_ty());
+ let ty = match self.infer_path(p, tgt_expr.into()) {
+ Some(ty) => ty,
+ None => {
+ if matches!(p, Path::Normal { mod_path, .. } if mod_path.is_ident()) {
+ self.push_diagnostic(InferenceDiagnostic::UnresolvedIdent {
+ expr: tgt_expr,
+ });
+ }
+ self.err_ty()
+ }
+ };
self.resolver.reset_to_guard(g);
ty
}
diff --git a/crates/hir-ty/src/mir/eval/shim.rs b/crates/hir-ty/src/mir/eval/shim.rs
index fbe6a982d6..628a1fe2d2 100644
--- a/crates/hir-ty/src/mir/eval/shim.rs
+++ b/crates/hir-ty/src/mir/eval/shim.rs
@@ -1,12 +1,20 @@
//! Interpret intrinsics, lang items and `extern "C"` wellknown functions which their implementation
//! is not available.
-
+//!
use std::cmp;
use chalk_ir::TyKind;
-use hir_def::builtin_type::{BuiltinInt, BuiltinUint};
+use hir_def::{
+ builtin_type::{BuiltinInt, BuiltinUint},
+ resolver::HasResolver,
+};
-use super::*;
+use crate::mir::eval::{
+ name, pad16, static_lifetime, Address, AdtId, Arc, BuiltinType, Evaluator, FunctionId,
+ HasModule, HirDisplay, Interned, InternedClosure, Interner, Interval, IntervalAndTy,
+ IntervalOrOwned, ItemContainerId, LangItem, Layout, Locals, Lookup, MirEvalError, MirSpan,
+ ModPath, Mutability, Result, Substitution, Ty, TyBuilder, TyExt,
+};
mod simd;
diff --git a/crates/hir-ty/src/mir/eval/shim/simd.rs b/crates/hir-ty/src/mir/eval/shim/simd.rs
index eddfd0acfb..e229a4ab31 100644
--- a/crates/hir-ty/src/mir/eval/shim/simd.rs
+++ b/crates/hir-ty/src/mir/eval/shim/simd.rs
@@ -2,6 +2,7 @@
use std::cmp::Ordering;
+use crate::consteval::try_const_usize;
use crate::TyKind;
use super::*;
diff --git a/crates/hir-ty/src/mir/lower.rs b/crates/hir-ty/src/mir/lower.rs
index 9fe3d5b77a..ed316f9726 100644
--- a/crates/hir-ty/src/mir/lower.rs
+++ b/crates/hir-ty/src/mir/lower.rs
@@ -1,6 +1,6 @@
//! This module generates a polymorphic MIR from a hir body
-use std::{fmt::Write, mem};
+use std::{fmt::Write, iter, mem};
use base_db::{salsa::Cycle, FileId};
use chalk_ir::{BoundVar, ConstData, DebruijnIndex, TyKind};
@@ -14,27 +14,37 @@ use hir_def::{
lang_item::{LangItem, LangItemTarget},
path::Path,
resolver::{resolver_for_expr, HasResolver, ResolveValueResult, ValueNs},
- AdtId, EnumVariantId, GeneralConstId, HasModule, ItemContainerId, LocalFieldId,
+ AdtId, DefWithBodyId, EnumVariantId, GeneralConstId, HasModule, ItemContainerId, LocalFieldId,
Lookup, TraitId, TupleId, TypeOrConstParamId,
};
use hir_expand::name::Name;
+use la_arena::ArenaMap;
+use rustc_hash::FxHashMap;
use syntax::TextRange;
use triomphe::Arc;
use crate::{
consteval::ConstEvalError,
- db::InternedClosure,
+ db::{HirDatabase, InternedClosure},
+ display::HirDisplay,
infer::{CaptureKind, CapturedItem, TypeMismatch},
inhabitedness::is_ty_uninhabited_from,
layout::LayoutError,
+ mapping::ToChalk,
+ mir::{
+ intern_const_scalar, return_slot, AggregateKind, Arena, BasicBlock, BasicBlockId, BinOp,
+ BorrowKind, CastKind, ClosureId, ConstScalar, Either, Expr, FieldId, Idx, InferenceResult,
+ Interner, Local, LocalId, MemoryMap, MirBody, MirSpan, Mutability, Operand, Place,
+ PlaceElem, PointerCast, ProjectionElem, ProjectionStore, RawIdx, Rvalue, Statement,
+ StatementKind, Substitution, SwitchTargets, Terminator, TerminatorKind, TupleFieldId, Ty,
+ UnOp, VariantId,
+ },
static_lifetime,
traits::FnTrait,
utils::{generics, ClosureSubst},
Adjust, Adjustment, AutoBorrow, CallableDefId, TyBuilder, TyExt,
};
-use super::*;
-
mod as_place;
mod pattern_matching;
diff --git a/crates/hir-ty/src/mir/lower/pattern_matching.rs b/crates/hir-ty/src/mir/lower/pattern_matching.rs
index 02b1494062..85c8d1685b 100644
--- a/crates/hir-ty/src/mir/lower/pattern_matching.rs
+++ b/crates/hir-ty/src/mir/lower/pattern_matching.rs
@@ -1,10 +1,17 @@
//! MIR lowering for patterns
-use hir_def::AssocItemId;
+use hir_def::{hir::LiteralOrConst, resolver::HasResolver, AssocItemId};
-use crate::BindingMode;
-
-use super::*;
+use crate::{
+ mir::lower::{
+ BasicBlockId, BinOp, BindingId, BorrowKind, Either, Expr, FieldId, Idx, Interner,
+ MemoryMap, MirLowerCtx, MirLowerError, MirSpan, Mutability, Operand, Pat, PatId, Place,
+ PlaceElem, ProjectionElem, RecordFieldPat, ResolveValueResult, Result, Rvalue,
+ Substitution, SwitchTargets, TerminatorKind, TupleFieldId, TupleId, TyBuilder, TyKind,
+ ValueNs, VariantData, VariantId,
+ },
+ BindingMode,
+};
macro_rules! not_supported {
($x: expr) => {
diff --git a/crates/hir-ty/src/tests/traits.rs b/crates/hir-ty/src/tests/traits.rs
index db14addaf1..879c69c758 100644
--- a/crates/hir-ty/src/tests/traits.rs
+++ b/crates/hir-ty/src/tests/traits.rs
@@ -4553,3 +4553,58 @@ fn foo() {
"#,
);
}
+
+#[test]
+fn auto_trait_bound() {
+ check_types(
+ r#"
+//- minicore: sized
+auto trait Send {}
+impl<T> !Send for *const T {}
+
+struct Yes;
+trait IsSend { const IS_SEND: Yes; }
+impl<T: Send> IsSend for T { const IS_SEND: Yes = Yes; }
+
+struct Struct<T>(T);
+enum Enum<T> { A, B(T) }
+union Union<T> { t: T }
+
+#[lang = "phantom_data"]
+struct PhantomData<T: ?Sized>;
+
+fn f<T: Send, U>() {
+ T::IS_SEND;
+ //^^^^^^^^^^Yes
+ U::IS_SEND;
+ //^^^^^^^^^^{unknown}
+ <*const T>::IS_SEND;
+ //^^^^^^^^^^^^^^^^^^^{unknown}
+ Struct::<T>::IS_SEND;
+ //^^^^^^^^^^^^^^^^^^^^Yes
+ Struct::<U>::IS_SEND;
+ //^^^^^^^^^^^^^^^^^^^^Yes
+ Struct::<*const T>::IS_SEND;
+ //^^^^^^^^^^^^^^^^^^^^^^^^^^^Yes
+ Enum::<T>::IS_SEND;
+ //^^^^^^^^^^^^^^^^^^Yes
+ Enum::<U>::IS_SEND;
+ //^^^^^^^^^^^^^^^^^^Yes
+ Enum::<*const T>::IS_SEND;
+ //^^^^^^^^^^^^^^^^^^^^^^^^^Yes
+ Union::<T>::IS_SEND;
+ //^^^^^^^^^^^^^^^^^^^Yes
+ Union::<U>::IS_SEND;
+ //^^^^^^^^^^^^^^^^^^^Yes
+ Union::<*const T>::IS_SEND;
+ //^^^^^^^^^^^^^^^^^^^^^^^^^^Yes
+ PhantomData::<T>::IS_SEND;
+ //^^^^^^^^^^^^^^^^^^^^^^^^^Yes
+ PhantomData::<U>::IS_SEND;
+ //^^^^^^^^^^^^^^^^^^^^^^^^^{unknown}
+ PhantomData::<*const T>::IS_SEND;
+ //^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^{unknown}
+}
+"#,
+ );
+}
diff --git a/crates/hir/src/diagnostics.rs b/crates/hir/src/diagnostics.rs
index 08843a6c99..80cd0c9c79 100644
--- a/crates/hir/src/diagnostics.rs
+++ b/crates/hir/src/diagnostics.rs
@@ -64,6 +64,7 @@ diagnostics![
MissingUnsafe,
MovedOutOfRef,
NeedMut,
+ NonExhaustiveLet,
NoSuchField,
PrivateAssocItem,
PrivateField,
@@ -86,6 +87,7 @@ diagnostics![
UnresolvedMacroCall,
UnresolvedMethodCall,
UnresolvedModule,
+ UnresolvedIdent,
UnresolvedProcMacro,
UnusedMut,
UnusedVariable,
@@ -242,6 +244,11 @@ pub struct UnresolvedAssocItem {
}
#[derive(Debug)]
+pub struct UnresolvedIdent {
+ pub expr: InFile<AstPtr<ast::Expr>>,
+}
+
+#[derive(Debug)]
pub struct PrivateField {
pub expr: InFile<AstPtr<ast::Expr>>,
pub field: Field,
@@ -281,6 +288,12 @@ pub struct MissingMatchArms {
}
#[derive(Debug)]
+pub struct NonExhaustiveLet {
+ pub pat: InFile<AstPtr<ast::Pat>>,
+ pub uncovered_patterns: String,
+}
+
+#[derive(Debug)]
pub struct TypeMismatch {
pub expr_or_pat: InFile<AstPtr<Either<ast::Expr, ast::Pat>>>,
pub expected: Type,
@@ -456,6 +469,22 @@ impl AnyDiagnostic {
Err(SyntheticSyntax) => (),
}
}
+ BodyValidationDiagnostic::NonExhaustiveLet { pat, uncovered_patterns } => {
+ match source_map.pat_syntax(pat) {
+ Ok(source_ptr) => {
+ if let Some(ast_pat) = source_ptr.value.cast::<ast::Pat>() {
+ return Some(
+ NonExhaustiveLet {
+ pat: InFile::new(source_ptr.file_id, ast_pat),
+ uncovered_patterns,
+ }
+ .into(),
+ );
+ }
+ }
+ Err(SyntheticSyntax) => {}
+ }
+ }
BodyValidationDiagnostic::RemoveTrailingReturn { return_expr } => {
if let Ok(source_ptr) = source_map.expr_syntax(return_expr) {
// Filters out desugared return expressions (e.g. desugared try operators).
@@ -565,6 +594,10 @@ impl AnyDiagnostic {
};
UnresolvedAssocItem { expr_or_pat }.into()
}
+ &InferenceDiagnostic::UnresolvedIdent { expr } => {
+ let expr = expr_syntax(expr);
+ UnresolvedIdent { expr }.into()
+ }
&InferenceDiagnostic::BreakOutsideOfLoop { expr, is_break, bad_value_break } => {
let expr = expr_syntax(expr);
BreakOutsideOfLoop { expr, is_break, bad_value_break }.into()
diff --git a/crates/hir/src/lib.rs b/crates/hir/src/lib.rs
index 08f7bb14ca..2d8811cf5e 100644
--- a/crates/hir/src/lib.rs
+++ b/crates/hir/src/lib.rs
@@ -2653,6 +2653,37 @@ impl ItemInNs {
}
}
+/// Invariant: `inner.as_extern_assoc_item(db).is_some()`
+/// We do not actively enforce this invariant.
+#[derive(Debug, Copy, Clone, PartialEq, Eq, Hash)]
+pub enum ExternAssocItem {
+ Function(Function),
+ Static(Static),
+ TypeAlias(TypeAlias),
+}
+
+pub trait AsExternAssocItem {
+ fn as_extern_assoc_item(self, db: &dyn HirDatabase) -> Option<ExternAssocItem>;
+}
+
+impl AsExternAssocItem for Function {
+ fn as_extern_assoc_item(self, db: &dyn HirDatabase) -> Option<ExternAssocItem> {
+ as_extern_assoc_item(db, ExternAssocItem::Function, self.id)
+ }
+}
+
+impl AsExternAssocItem for Static {
+ fn as_extern_assoc_item(self, db: &dyn HirDatabase) -> Option<ExternAssocItem> {
+ as_extern_assoc_item(db, ExternAssocItem::Static, self.id)
+ }
+}
+
+impl AsExternAssocItem for TypeAlias {
+ fn as_extern_assoc_item(self, db: &dyn HirDatabase) -> Option<ExternAssocItem> {
+ as_extern_assoc_item(db, ExternAssocItem::TypeAlias, self.id)
+ }
+}
+
/// Invariant: `inner.as_assoc_item(db).is_some()`
/// We do not actively enforce this invariant.
#[derive(Debug, Copy, Clone, PartialEq, Eq, Hash)]
@@ -2727,6 +2758,63 @@ where
}
}
+fn as_extern_assoc_item<'db, ID, DEF, LOC>(
+ db: &(dyn HirDatabase + 'db),
+ ctor: impl FnOnce(DEF) -> ExternAssocItem,
+ id: ID,
+) -> Option<ExternAssocItem>
+where
+ ID: Lookup<Database<'db> = dyn DefDatabase + 'db, Data = AssocItemLoc<LOC>>,
+ DEF: From<ID>,
+ LOC: ItemTreeNode,
+{
+ match id.lookup(db.upcast()).container {
+ ItemContainerId::ExternBlockId(_) => Some(ctor(DEF::from(id))),
+ ItemContainerId::TraitId(_) | ItemContainerId::ImplId(_) | ItemContainerId::ModuleId(_) => {
+ None
+ }
+ }
+}
+
+impl ExternAssocItem {
+ pub fn name(self, db: &dyn HirDatabase) -> Name {
+ match self {
+ Self::Function(it) => it.name(db),
+ Self::Static(it) => it.name(db),
+ Self::TypeAlias(it) => it.name(db),
+ }
+ }
+
+ pub fn module(self, db: &dyn HirDatabase) -> Module {
+ match self {
+ Self::Function(f) => f.module(db),
+ Self::Static(c) => c.module(db),
+ Self::TypeAlias(t) => t.module(db),
+ }
+ }
+
+ pub fn as_function(self) -> Option<Function> {
+ match self {
+ Self::Function(v) => Some(v),
+ _ => None,
+ }
+ }
+
+ pub fn as_static(self) -> Option<Static> {
+ match self {
+ Self::Static(v) => Some(v),
+ _ => None,
+ }
+ }
+
+ pub fn as_type_alias(self) -> Option<TypeAlias> {
+ match self {
+ Self::TypeAlias(v) => Some(v),
+ _ => None,
+ }
+ }
+}
+
impl AssocItem {
pub fn name(self, db: &dyn HirDatabase) -> Option<Name> {
match self {
@@ -3798,9 +3886,9 @@ impl Type {
// For non-phantom_data adts we check variants/fields as well as generic parameters
TyKind::Adt(adt_id, substitution)
- if !db.struct_datum(krate, *adt_id).flags.phantom_data =>
+ if !db.adt_datum(krate, *adt_id).flags.phantom_data =>
{
- let adt_datum = &db.struct_datum(krate, *adt_id);
+ let adt_datum = &db.adt_datum(krate, *adt_id);
let adt_datum_bound =
adt_datum.binders.clone().substitute(Interner, substitution);
adt_datum_bound
diff --git a/crates/hir/src/term_search/tactics.rs b/crates/hir/src/term_search/tactics.rs
index 666d63ac15..edbf75affe 100644
--- a/crates/hir/src/term_search/tactics.rs
+++ b/crates/hir/src/term_search/tactics.rs
@@ -281,14 +281,14 @@ pub(super) fn type_constructor<'a, DB: HirDatabase>(
if ctx.config.enable_borrowcheck && struct_ty.contains_reference(db) {
return None;
}
- let fileds = it.fields(db);
+ let fields = it.fields(db);
// Check if all fields are visible, otherwise we cannot fill them
- if fileds.iter().any(|it| !it.is_visible_from(db, module)) {
+ if fields.iter().any(|it| !it.is_visible_from(db, module)) {
return None;
}
// Early exit if some param cannot be filled from lookup
- let param_exprs: Vec<Vec<Expr>> = fileds
+ let param_exprs: Vec<Vec<Expr>> = fields
.into_iter()
.map(|field| lookup.find(db, &field.ty(db)))
.collect::<Option<_>>()?;
diff --git a/crates/ide-completion/src/item.rs b/crates/ide-completion/src/item.rs
index c2c0641961..4bab288685 100644
--- a/crates/ide-completion/src/item.rs
+++ b/crates/ide-completion/src/item.rs
@@ -308,7 +308,7 @@ impl CompletionRelevance {
// When a fn is bumped due to return type:
// Bump Constructor or Builder methods with no arguments,
- // over them tha with self arguments
+ // over them than with self arguments
if fn_score > 0 {
if !asf.has_params {
// bump associated functions
diff --git a/crates/ide-db/src/apply_change.rs b/crates/ide-db/src/apply_change.rs
index 296253aa1e..2b2df144d6 100644
--- a/crates/ide-db/src/apply_change.rs
+++ b/crates/ide-db/src/apply_change.rs
@@ -17,7 +17,7 @@ impl RootDatabase {
pub fn request_cancellation(&mut self) {
let _p =
tracing::span!(tracing::Level::INFO, "RootDatabase::request_cancellation").entered();
- self.salsa_runtime_mut().synthetic_write(Durability::LOW);
+ self.synthetic_write(Durability::LOW);
}
pub fn apply_change(&mut self, change: Change) {
@@ -124,7 +124,7 @@ impl RootDatabase {
hir::db::InternCoroutineQuery
hir::db::AssociatedTyDataQuery
hir::db::TraitDatumQuery
- hir::db::StructDatumQuery
+ hir::db::AdtDatumQuery
hir::db::ImplDatumQuery
hir::db::FnDefDatumQuery
hir::db::FnDefVarianceQuery
diff --git a/crates/ide-db/src/defs.rs b/crates/ide-db/src/defs.rs
index d95d94ec72..1b6ff8bad5 100644
--- a/crates/ide-db/src/defs.rs
+++ b/crates/ide-db/src/defs.rs
@@ -8,11 +8,11 @@
use arrayvec::ArrayVec;
use either::Either;
use hir::{
- Adt, AsAssocItem, AssocItem, AttributeTemplate, BuiltinAttr, BuiltinType, Const, Crate,
- DefWithBody, DeriveHelper, DocLinkDef, ExternCrateDecl, Field, Function, GenericParam,
- HasVisibility, HirDisplay, Impl, Label, Local, Macro, Module, ModuleDef, Name, PathResolution,
- Semantics, Static, ToolModule, Trait, TraitAlias, TupleField, TypeAlias, Variant, VariantDef,
- Visibility,
+ Adt, AsAssocItem, AsExternAssocItem, AssocItem, AttributeTemplate, BuiltinAttr, BuiltinType,
+ Const, Crate, DefWithBody, DeriveHelper, DocLinkDef, ExternAssocItem, ExternCrateDecl, Field,
+ Function, GenericParam, HasVisibility, HirDisplay, Impl, Label, Local, Macro, Module,
+ ModuleDef, Name, PathResolution, Semantics, Static, ToolModule, Trait, TraitAlias, TupleField,
+ TypeAlias, Variant, VariantDef, Visibility,
};
use stdx::{format_to, impl_from};
use syntax::{
@@ -213,8 +213,8 @@ impl Definition {
})
}
- pub fn label(&self, db: &RootDatabase) -> Option<String> {
- let label = match *self {
+ pub fn label(&self, db: &RootDatabase) -> String {
+ match *self {
Definition::Macro(it) => it.display(db).to_string(),
Definition::Field(it) => it.display(db).to_string(),
Definition::TupleField(it) => it.display(db).to_string(),
@@ -241,7 +241,11 @@ impl Definition {
}
}
Definition::SelfType(impl_def) => {
- impl_def.self_ty(db).as_adt().and_then(|adt| Definition::Adt(adt).label(db))?
+ let self_ty = &impl_def.self_ty(db);
+ match self_ty.as_adt() {
+ Some(it) => it.display(db).to_string(),
+ None => self_ty.display(db).to_string(),
+ }
}
Definition::GenericParam(it) => it.display(db).to_string(),
Definition::Label(it) => it.name(db).display(db).to_string(),
@@ -249,8 +253,7 @@ impl Definition {
Definition::BuiltinAttr(it) => format!("#[{}]", it.name(db)),
Definition::ToolModule(it) => it.name(db).to_string(),
Definition::DeriveHelper(it) => format!("derive_helper {}", it.name(db).display(db)),
- };
- Some(label)
+ }
}
}
@@ -739,6 +742,17 @@ impl AsAssocItem for Definition {
}
}
+impl AsExternAssocItem for Definition {
+ fn as_extern_assoc_item(self, db: &dyn hir::db::HirDatabase) -> Option<ExternAssocItem> {
+ match self {
+ Definition::Function(it) => it.as_extern_assoc_item(db),
+ Definition::Static(it) => it.as_extern_assoc_item(db),
+ Definition::TypeAlias(it) => it.as_extern_assoc_item(db),
+ _ => None,
+ }
+ }
+}
+
impl From<AssocItem> for Definition {
fn from(assoc_item: AssocItem) -> Self {
match assoc_item {
diff --git a/crates/ide-db/src/imports/insert_use/tests.rs b/crates/ide-db/src/imports/insert_use/tests.rs
index 6b0fecae26..10c285a13f 100644
--- a/crates/ide-db/src/imports/insert_use/tests.rs
+++ b/crates/ide-db/src/imports/insert_use/tests.rs
@@ -1,4 +1,3 @@
-use hir::PrefixKind;
use stdx::trim_indent;
use test_fixture::WithFixture;
use test_utils::{assert_eq_text, CURSOR_MARKER};
diff --git a/crates/ide-db/src/lib.rs b/crates/ide-db/src/lib.rs
index 2881748dd4..d31dad514a 100644
--- a/crates/ide-db/src/lib.rs
+++ b/crates/ide-db/src/lib.rs
@@ -280,7 +280,7 @@ impl RootDatabase {
// hir_db::InternCoroutineQuery
hir_db::AssociatedTyDataQuery
hir_db::TraitDatumQuery
- hir_db::StructDatumQuery
+ hir_db::AdtDatumQuery
hir_db::ImplDatumQuery
hir_db::FnDefDatumQuery
hir_db::FnDefVarianceQuery
diff --git a/crates/ide-db/src/symbol_index.rs b/crates/ide-db/src/symbol_index.rs
index 722161282f..c65467a432 100644
--- a/crates/ide-db/src/symbol_index.rs
+++ b/crates/ide-db/src/symbol_index.rs
@@ -394,7 +394,6 @@ impl Query {
mod tests {
use expect_test::expect_file;
- use hir::symbols::SymbolCollector;
use test_fixture::WithFixture;
use super::*;
diff --git a/crates/ide-diagnostics/src/handlers/inactive_code.rs b/crates/ide-diagnostics/src/handlers/inactive_code.rs
index 7db5ea04fb..785a42352b 100644
--- a/crates/ide-diagnostics/src/handlers/inactive_code.rs
+++ b/crates/ide-diagnostics/src/handlers/inactive_code.rs
@@ -60,6 +60,7 @@ fn f() {
#[cfg(a)] let x = 0; // let statement
//^^^^^^^^^^^^^^^^^^^^ weak: code is inactive due to #[cfg] directives: a is disabled
+ fn abc() {}
abc(#[cfg(a)] 0);
//^^^^^^^^^^^ weak: code is inactive due to #[cfg] directives: a is disabled
let x = Struct {
diff --git a/crates/ide-diagnostics/src/handlers/incorrect_case.rs b/crates/ide-diagnostics/src/handlers/incorrect_case.rs
index 5e2541795c..db28928a24 100644
--- a/crates/ide-diagnostics/src/handlers/incorrect_case.rs
+++ b/crates/ide-diagnostics/src/handlers/incorrect_case.rs
@@ -512,7 +512,7 @@ impl BAD_TRAIT for () {
fn BadFunction() {}
}
"#,
- std::iter::once("unused_variables".to_owned()),
+ &["unused_variables"],
);
}
diff --git a/crates/ide-diagnostics/src/handlers/missing_fields.rs b/crates/ide-diagnostics/src/handlers/missing_fields.rs
index c70f39eb28..09daefd084 100644
--- a/crates/ide-diagnostics/src/handlers/missing_fields.rs
+++ b/crates/ide-diagnostics/src/handlers/missing_fields.rs
@@ -634,7 +634,8 @@ struct TestStruct { one: i32, two: i64 }
fn test_fn() {
let one = 1;
- let s = TestStruct{ ..a };
+ let a = TestStruct{ one, two: 2 };
+ let _ = TestStruct{ ..a };
}
"#,
);
diff --git a/crates/ide-diagnostics/src/handlers/missing_match_arms.rs b/crates/ide-diagnostics/src/handlers/missing_match_arms.rs
index 7632fdf1d0..8596f5792e 100644
--- a/crates/ide-diagnostics/src/handlers/missing_match_arms.rs
+++ b/crates/ide-diagnostics/src/handlers/missing_match_arms.rs
@@ -18,7 +18,9 @@ pub(crate) fn missing_match_arms(
#[cfg(test)]
mod tests {
use crate::{
- tests::{check_diagnostics, check_diagnostics_with_config},
+ tests::{
+ check_diagnostics, check_diagnostics_with_config, check_diagnostics_with_disabled,
+ },
DiagnosticsConfig,
};
@@ -282,7 +284,7 @@ fn main() {
cov_mark::check_count!(validate_match_bailed_out, 4);
// Match statements with arms that don't match the
// expression pattern do not fire this diagnostic.
- check_diagnostics(
+ check_diagnostics_with_disabled(
r#"
enum Either { A, B }
enum Either2 { C, D }
@@ -307,6 +309,7 @@ fn main() {
match Unresolved::Bar { Unresolved::Baz => () }
}
"#,
+ &["E0425"],
);
}
@@ -397,11 +400,11 @@ fn main() {
match loop {} {
Either::A => (),
}
- match loop { break Foo::A } {
- //^^^^^^^^^^^^^^^^^^^^^ error: missing match arm: `B` not covered
+ match loop { break Either::A } {
+ //^^^^^^^^^^^^^^^^^^^^^^^^ error: missing match arm: `B` not covered
Either::A => (),
}
- match loop { break Foo::A } {
+ match loop { break Either::A } {
Either::A => (),
Either::B => (),
}
@@ -977,7 +980,7 @@ fn f(ty: Enum) {
#[test]
fn unexpected_ty_fndef() {
cov_mark::check!(validate_match_bailed_out);
- check_diagnostics(
+ check_diagnostics_with_disabled(
r"
enum Exp {
Tuple(()),
@@ -987,6 +990,7 @@ fn f() {
Exp::Tuple => {}
}
}",
+ &["E0425"],
);
}
diff --git a/crates/ide-diagnostics/src/handlers/mutability_errors.rs b/crates/ide-diagnostics/src/handlers/mutability_errors.rs
index bdb55a9d98..91f1058d65 100644
--- a/crates/ide-diagnostics/src/handlers/mutability_errors.rs
+++ b/crates/ide-diagnostics/src/handlers/mutability_errors.rs
@@ -448,7 +448,7 @@ fn main(b: bool) {
&mut x;
}
"#,
- std::iter::once("remove-unnecessary-else".to_owned()),
+ &["remove-unnecessary-else"],
);
check_diagnostics_with_disabled(
r#"
@@ -463,7 +463,7 @@ fn main(b: bool) {
&mut x;
}
"#,
- std::iter::once("remove-unnecessary-else".to_owned()),
+ &["remove-unnecessary-else"],
);
}
@@ -817,7 +817,7 @@ fn f() {
//- minicore: option
fn f(_: i32) {}
fn main() {
- let ((Some(mut x), None) | (_, Some(mut x))) = (None, Some(7));
+ let ((Some(mut x), None) | (_, Some(mut x))) = (None, Some(7)) else { return };
//^^^^^ 💡 warn: variable does not need to be mutable
f(x);
}
diff --git a/crates/ide-diagnostics/src/handlers/non_exhaustive_let.rs b/crates/ide-diagnostics/src/handlers/non_exhaustive_let.rs
new file mode 100644
index 0000000000..1a4d2877ef
--- /dev/null
+++ b/crates/ide-diagnostics/src/handlers/non_exhaustive_let.rs
@@ -0,0 +1,47 @@
+use crate::{Diagnostic, DiagnosticCode, DiagnosticsContext};
+
+// Diagnostic: non-exhaustive-let
+//
+// This diagnostic is triggered if a `let` statement without an `else` branch has a non-exhaustive
+// pattern.
+pub(crate) fn non_exhaustive_let(
+ ctx: &DiagnosticsContext<'_>,
+ d: &hir::NonExhaustiveLet,
+) -> Diagnostic {
+ Diagnostic::new_with_syntax_node_ptr(
+ ctx,
+ DiagnosticCode::RustcHardError("E0005"),
+ format!("non-exhaustive pattern: {}", d.uncovered_patterns),
+ d.pat.map(Into::into),
+ )
+}
+
+#[cfg(test)]
+mod tests {
+ use crate::tests::check_diagnostics;
+
+ #[test]
+ fn option_nonexhaustive() {
+ check_diagnostics(
+ r#"
+//- minicore: option
+fn main() {
+ let None = Some(5);
+ //^^^^ error: non-exhaustive pattern: `Some(_)` not covered
+}
+"#,
+ );
+ }
+
+ #[test]
+ fn option_exhaustive() {
+ check_diagnostics(
+ r#"
+//- minicore: option
+fn main() {
+ let Some(_) | None = Some(5);
+}
+"#,
+ );
+ }
+}
diff --git a/crates/ide-diagnostics/src/handlers/remove_trailing_return.rs b/crates/ide-diagnostics/src/handlers/remove_trailing_return.rs
index b7667dc318..7a040e46e3 100644
--- a/crates/ide-diagnostics/src/handlers/remove_trailing_return.rs
+++ b/crates/ide-diagnostics/src/handlers/remove_trailing_return.rs
@@ -140,7 +140,7 @@ fn foo(x: usize) -> u8 {
} //^^^^^^^^^ 💡 weak: replace return <expr>; with <expr>
}
"#,
- std::iter::once("remove-unnecessary-else".to_owned()),
+ &["remove-unnecessary-else"],
);
}
diff --git a/crates/ide-diagnostics/src/handlers/remove_unnecessary_else.rs b/crates/ide-diagnostics/src/handlers/remove_unnecessary_else.rs
index ae8241ec2c..47844876dc 100644
--- a/crates/ide-diagnostics/src/handlers/remove_unnecessary_else.rs
+++ b/crates/ide-diagnostics/src/handlers/remove_unnecessary_else.rs
@@ -2,7 +2,10 @@ use hir::{db::ExpandDatabase, diagnostics::RemoveUnnecessaryElse, HirFileIdExt};
use ide_db::{assists::Assist, source_change::SourceChange};
use itertools::Itertools;
use syntax::{
- ast::{self, edit::IndentLevel},
+ ast::{
+ self,
+ edit::{AstNodeEdit, IndentLevel},
+ },
AstNode, SyntaxToken, TextRange,
};
use text_edit::TextEdit;
@@ -41,10 +44,15 @@ fn fixes(ctx: &DiagnosticsContext<'_>, d: &RemoveUnnecessaryElse) -> Option<Vec<
indent = indent + 1;
}
let else_replacement = match if_expr.else_branch()? {
- ast::ElseBranch::Block(ref block) => {
- block.statements().map(|stmt| format!("\n{indent}{stmt}")).join("")
- }
- ast::ElseBranch::IfExpr(ref nested_if_expr) => {
+ ast::ElseBranch::Block(block) => block
+ .statements()
+ .map(|stmt| format!("\n{indent}{stmt}"))
+ .chain(block.tail_expr().map(|tail| format!("\n{indent}{tail}")))
+ .join(""),
+ ast::ElseBranch::IfExpr(mut nested_if_expr) => {
+ if has_parent_if_expr {
+ nested_if_expr = nested_if_expr.indent(IndentLevel(1))
+ }
format!("\n{indent}{nested_if_expr}")
}
};
@@ -87,15 +95,11 @@ fn fixes(ctx: &DiagnosticsContext<'_>, d: &RemoveUnnecessaryElse) -> Option<Vec<
#[cfg(test)]
mod tests {
- use crate::tests::{check_diagnostics, check_diagnostics_with_disabled, check_fix};
-
- fn check_diagnostics_with_needless_return_disabled(ra_fixture: &str) {
- check_diagnostics_with_disabled(ra_fixture, std::iter::once("needless_return".to_owned()));
- }
+ use crate::tests::{check_diagnostics_with_disabled, check_fix};
#[test]
fn remove_unnecessary_else_for_return() {
- check_diagnostics_with_needless_return_disabled(
+ check_diagnostics_with_disabled(
r#"
fn test() {
if foo {
@@ -106,6 +110,7 @@ fn test() {
}
}
"#,
+ &["needless_return", "E0425"],
);
check_fix(
r#"
@@ -130,7 +135,7 @@ fn test() {
#[test]
fn remove_unnecessary_else_for_return2() {
- check_diagnostics_with_needless_return_disabled(
+ check_diagnostics_with_disabled(
r#"
fn test() {
if foo {
@@ -143,6 +148,7 @@ fn test() {
}
}
"#,
+ &["needless_return", "E0425"],
);
check_fix(
r#"
@@ -172,8 +178,44 @@ fn test() {
}
#[test]
+ fn remove_unnecessary_else_for_return3() {
+ check_diagnostics_with_disabled(
+ r#"
+fn test(a: bool) -> i32 {
+ if a {
+ return 1;
+ } else {
+ //^^^^ 💡 weak: remove unnecessary else block
+ 0
+ }
+}
+"#,
+ &["needless_return", "E0425"],
+ );
+ check_fix(
+ r#"
+fn test(a: bool) -> i32 {
+ if a {
+ return 1;
+ } else$0 {
+ 0
+ }
+}
+"#,
+ r#"
+fn test(a: bool) -> i32 {
+ if a {
+ return 1;
+ }
+ 0
+}
+"#,
+ );
+ }
+
+ #[test]
fn remove_unnecessary_else_for_return_in_child_if_expr() {
- check_diagnostics_with_needless_return_disabled(
+ check_diagnostics_with_disabled(
r#"
fn test() {
if foo {
@@ -186,6 +228,7 @@ fn test() {
}
}
"#,
+ &["needless_return", "E0425"],
);
check_fix(
r#"
@@ -215,8 +258,43 @@ fn test() {
}
#[test]
+ fn remove_unnecessary_else_for_return_in_child_if_expr2() {
+ check_fix(
+ r#"
+fn test() {
+ if foo {
+ do_something();
+ } else if qux {
+ return bar;
+ } else$0 if quux {
+ do_something_else();
+ } else {
+ do_something_else2();
+ }
+}
+"#,
+ r#"
+fn test() {
+ if foo {
+ do_something();
+ } else {
+ if qux {
+ return bar;
+ }
+ if quux {
+ do_something_else();
+ } else {
+ do_something_else2();
+ }
+ }
+}
+"#,
+ );
+ }
+
+ #[test]
fn remove_unnecessary_else_for_break() {
- check_diagnostics(
+ check_diagnostics_with_disabled(
r#"
fn test() {
loop {
@@ -229,6 +307,7 @@ fn test() {
}
}
"#,
+ &["E0425"],
);
check_fix(
r#"
@@ -257,7 +336,7 @@ fn test() {
#[test]
fn remove_unnecessary_else_for_continue() {
- check_diagnostics(
+ check_diagnostics_with_disabled(
r#"
fn test() {
loop {
@@ -270,6 +349,7 @@ fn test() {
}
}
"#,
+ &["E0425"],
);
check_fix(
r#"
@@ -298,7 +378,7 @@ fn test() {
#[test]
fn remove_unnecessary_else_for_never() {
- check_diagnostics(
+ check_diagnostics_with_disabled(
r#"
fn test() {
if foo {
@@ -313,6 +393,7 @@ fn never() -> ! {
loop {}
}
"#,
+ &["E0425"],
);
check_fix(
r#"
@@ -345,7 +426,7 @@ fn never() -> ! {
#[test]
fn no_diagnostic_if_no_else_branch() {
- check_diagnostics(
+ check_diagnostics_with_disabled(
r#"
fn test() {
if foo {
@@ -355,12 +436,13 @@ fn test() {
do_something_else();
}
"#,
+ &["E0425"],
);
}
#[test]
fn no_diagnostic_if_no_divergence() {
- check_diagnostics(
+ check_diagnostics_with_disabled(
r#"
fn test() {
if foo {
@@ -370,12 +452,13 @@ fn test() {
}
}
"#,
+ &["E0425"],
);
}
#[test]
fn no_diagnostic_if_no_divergence_in_else_branch() {
- check_diagnostics_with_needless_return_disabled(
+ check_diagnostics_with_disabled(
r#"
fn test() {
if foo {
@@ -385,6 +468,43 @@ fn test() {
}
}
"#,
+ &["needless_return", "E0425"],
+ );
+ }
+
+ #[test]
+ fn no_diagnostic_if_not_expr_stmt() {
+ check_diagnostics_with_disabled(
+ r#"
+fn test1() {
+ let _x = if a {
+ return;
+ } else {
+ 1
+ };
+}
+
+fn test2() {
+ let _x = if a {
+ return;
+ } else if b {
+ return;
+ } else if c {
+ 1
+ } else {
+ return;
+ };
+}
+"#,
+ &["needless_return", "E0425"],
+ );
+ check_diagnostics_with_disabled(
+ r#"
+fn test3() -> u8 {
+ foo(if a { return 1 } else { 0 })
+}
+"#,
+ &["E0425"],
);
}
}
diff --git a/crates/ide-diagnostics/src/handlers/type_mismatch.rs b/crates/ide-diagnostics/src/handlers/type_mismatch.rs
index 8c97281b78..4c25532228 100644
--- a/crates/ide-diagnostics/src/handlers/type_mismatch.rs
+++ b/crates/ide-diagnostics/src/handlers/type_mismatch.rs
@@ -730,7 +730,7 @@ fn f() -> i32 {
}
fn g() { return; }
"#,
- std::iter::once("needless_return".to_owned()),
+ &["needless_return"],
);
}
diff --git a/crates/ide-diagnostics/src/handlers/undeclared_label.rs b/crates/ide-diagnostics/src/handlers/undeclared_label.rs
index a6a0fdc655..97943b7e8b 100644
--- a/crates/ide-diagnostics/src/handlers/undeclared_label.rs
+++ b/crates/ide-diagnostics/src/handlers/undeclared_label.rs
@@ -38,10 +38,12 @@ fn foo() {
fn while_let_loop_with_label_in_condition() {
check_diagnostics(
r#"
+//- minicore: option
+
fn foo() {
let mut optional = Some(0);
- 'my_label: while let Some(a) = match optional {
+ 'my_label: while let Some(_) = match optional {
None => break 'my_label,
Some(val) => Some(val),
} {
@@ -59,8 +61,8 @@ fn foo() {
r#"
//- minicore: iterator
fn foo() {
- 'xxx: for _ in unknown {
- 'yyy: for _ in unknown {
+ 'xxx: for _ in [] {
+ 'yyy: for _ in [] {
break 'xxx;
continue 'yyy;
break 'zzz;
diff --git a/crates/ide-diagnostics/src/handlers/unresolved_field.rs b/crates/ide-diagnostics/src/handlers/unresolved_field.rs
index 65abfd8a29..4c01a2d155 100644
--- a/crates/ide-diagnostics/src/handlers/unresolved_field.rs
+++ b/crates/ide-diagnostics/src/handlers/unresolved_field.rs
@@ -78,7 +78,9 @@ fn method_fix(
#[cfg(test)]
mod tests {
use crate::{
- tests::{check_diagnostics, check_diagnostics_with_config},
+ tests::{
+ check_diagnostics, check_diagnostics_with_config, check_diagnostics_with_disabled,
+ },
DiagnosticsConfig,
};
@@ -148,7 +150,7 @@ fn foo() {
#[test]
fn no_diagnostic_on_unknown() {
- check_diagnostics(
+ check_diagnostics_with_disabled(
r#"
fn foo() {
x.foo;
@@ -156,6 +158,7 @@ fn foo() {
(&((x,),),).foo;
}
"#,
+ &["E0425"],
);
}
diff --git a/crates/ide-diagnostics/src/handlers/unresolved_ident.rs b/crates/ide-diagnostics/src/handlers/unresolved_ident.rs
new file mode 100644
index 0000000000..295c8a2c61
--- /dev/null
+++ b/crates/ide-diagnostics/src/handlers/unresolved_ident.rs
@@ -0,0 +1,46 @@
+use crate::{Diagnostic, DiagnosticCode, DiagnosticsContext};
+
+// Diagnostic: unresolved-ident
+//
+// This diagnostic is triggered if an expr-position ident is invalid.
+pub(crate) fn unresolved_ident(
+ ctx: &DiagnosticsContext<'_>,
+ d: &hir::UnresolvedIdent,
+) -> Diagnostic {
+ Diagnostic::new_with_syntax_node_ptr(
+ ctx,
+ DiagnosticCode::RustcHardError("E0425"),
+ "no such value in this scope",
+ d.expr.map(Into::into),
+ )
+ .experimental()
+}
+
+#[cfg(test)]
+mod tests {
+ use crate::tests::check_diagnostics;
+
+ #[test]
+ fn missing() {
+ check_diagnostics(
+ r#"
+fn main() {
+ let _ = x;
+ //^ error: no such value in this scope
+}
+"#,
+ );
+ }
+
+ #[test]
+ fn present() {
+ check_diagnostics(
+ r#"
+fn main() {
+ let x = 5;
+ let _ = x;
+}
+"#,
+ );
+ }
+}
diff --git a/crates/ide-diagnostics/src/handlers/unresolved_method.rs b/crates/ide-diagnostics/src/handlers/unresolved_method.rs
index 648d081898..0614fdc551 100644
--- a/crates/ide-diagnostics/src/handlers/unresolved_method.rs
+++ b/crates/ide-diagnostics/src/handlers/unresolved_method.rs
@@ -335,8 +335,8 @@ fn main() {
r#"
struct Foo { bar: i32 }
fn foo() {
- Foo { bar: i32 }.bar();
- // ^^^ error: no method `bar` on type `Foo`, but a field with a similar name exists
+ Foo { bar: 0 }.bar();
+ // ^^^ error: no method `bar` on type `Foo`, but a field with a similar name exists
}
"#,
);
diff --git a/crates/ide-diagnostics/src/handlers/useless_braces.rs b/crates/ide-diagnostics/src/handlers/useless_braces.rs
index 863a7ab783..79bcaa0a9c 100644
--- a/crates/ide-diagnostics/src/handlers/useless_braces.rs
+++ b/crates/ide-diagnostics/src/handlers/useless_braces.rs
@@ -4,7 +4,7 @@ use ide_db::{
source_change::SourceChange,
};
use itertools::Itertools;
-use syntax::{ast, AstNode, SyntaxNode};
+use syntax::{ast, AstNode, SyntaxNode, SyntaxNodePtr};
use text_edit::TextEdit;
use crate::{fix, Diagnostic, DiagnosticCode};
@@ -43,7 +43,7 @@ pub(crate) fn useless_braces(
"Unnecessary braces in use statement".to_owned(),
FileRange { file_id, range: use_range },
)
- .with_main_node(InFile::new(file_id.into(), node.clone()))
+ .with_main_node(InFile::new(file_id.into(), SyntaxNodePtr::new(node)))
.with_fixes(Some(vec![fix(
"remove_braces",
"Remove unnecessary braces",
diff --git a/crates/ide-diagnostics/src/lib.rs b/crates/ide-diagnostics/src/lib.rs
index 9d21bb4cd9..9f4368b04e 100644
--- a/crates/ide-diagnostics/src/lib.rs
+++ b/crates/ide-diagnostics/src/lib.rs
@@ -41,6 +41,7 @@ mod handlers {
pub(crate) mod moved_out_of_ref;
pub(crate) mod mutability_errors;
pub(crate) mod no_such_field;
+ pub(crate) mod non_exhaustive_let;
pub(crate) mod private_assoc_item;
pub(crate) mod private_field;
pub(crate) mod remove_trailing_return;
@@ -58,6 +59,7 @@ mod handlers {
pub(crate) mod unresolved_assoc_item;
pub(crate) mod unresolved_extern_crate;
pub(crate) mod unresolved_field;
+ pub(crate) mod unresolved_ident;
pub(crate) mod unresolved_import;
pub(crate) mod unresolved_macro_call;
pub(crate) mod unresolved_method;
@@ -140,7 +142,7 @@ pub struct Diagnostic {
pub experimental: bool,
pub fixes: Option<Vec<Assist>>,
// The node that will be affected by `#[allow]` and similar attributes.
- pub main_node: Option<InFile<SyntaxNode>>,
+ pub main_node: Option<InFile<SyntaxNodePtr>>,
}
impl Diagnostic {
@@ -172,9 +174,8 @@ impl Diagnostic {
message: impl Into<String>,
node: InFile<SyntaxNodePtr>,
) -> Diagnostic {
- let file_id = node.file_id;
Diagnostic::new(code, message, ctx.sema.diagnostics_display_range(node))
- .with_main_node(node.map(|x| x.to_node(&ctx.sema.parse_or_expand(file_id))))
+ .with_main_node(node)
}
fn experimental(mut self) -> Diagnostic {
@@ -182,7 +183,7 @@ impl Diagnostic {
self
}
- fn with_main_node(mut self, main_node: InFile<SyntaxNode>) -> Diagnostic {
+ fn with_main_node(mut self, main_node: InFile<SyntaxNodePtr>) -> Diagnostic {
self.main_node = Some(main_node);
self
}
@@ -359,6 +360,7 @@ pub fn diagnostics(
AnyDiagnostic::MissingUnsafe(d) => handlers::missing_unsafe::missing_unsafe(&ctx, &d),
AnyDiagnostic::MovedOutOfRef(d) => handlers::moved_out_of_ref::moved_out_of_ref(&ctx, &d),
AnyDiagnostic::NeedMut(d) => handlers::mutability_errors::need_mut(&ctx, &d),
+ AnyDiagnostic::NonExhaustiveLet(d) => handlers::non_exhaustive_let::non_exhaustive_let(&ctx, &d),
AnyDiagnostic::NoSuchField(d) => handlers::no_such_field::no_such_field(&ctx, &d),
AnyDiagnostic::PrivateAssocItem(d) => handlers::private_assoc_item::private_assoc_item(&ctx, &d),
AnyDiagnostic::PrivateField(d) => handlers::private_field::private_field(&ctx, &d),
@@ -375,6 +377,7 @@ pub fn diagnostics(
AnyDiagnostic::UnresolvedAssocItem(d) => handlers::unresolved_assoc_item::unresolved_assoc_item(&ctx, &d),
AnyDiagnostic::UnresolvedExternCrate(d) => handlers::unresolved_extern_crate::unresolved_extern_crate(&ctx, &d),
AnyDiagnostic::UnresolvedField(d) => handlers::unresolved_field::unresolved_field(&ctx, &d),
+ AnyDiagnostic::UnresolvedIdent(d) => handlers::unresolved_ident::unresolved_ident(&ctx, &d),
AnyDiagnostic::UnresolvedImport(d) => handlers::unresolved_import::unresolved_import(&ctx, &d),
AnyDiagnostic::UnresolvedMacroCall(d) => handlers::unresolved_macro_call::unresolved_macro_call(&ctx, &d),
AnyDiagnostic::UnresolvedMethodCall(d) => handlers::unresolved_method::unresolved_method(&ctx, &d),
@@ -390,8 +393,17 @@ pub fn diagnostics(
res.push(d)
}
- let mut diagnostics_of_range =
- res.iter_mut().filter_map(|x| Some((x.main_node.clone()?, x))).collect::<FxHashMap<_, _>>();
+ let mut diagnostics_of_range = res
+ .iter_mut()
+ .filter_map(|it| {
+ Some((
+ it.main_node
+ .map(|ptr| ptr.map(|node| node.to_node(&ctx.sema.parse_or_expand(ptr.file_id))))
+ .clone()?,
+ it,
+ ))
+ })
+ .collect::<FxHashMap<_, _>>();
let mut rustc_stack: FxHashMap<String, Vec<Severity>> = FxHashMap::default();
let mut clippy_stack: FxHashMap<String, Vec<Severity>> = FxHashMap::default();
diff --git a/crates/ide-diagnostics/src/tests.rs b/crates/ide-diagnostics/src/tests.rs
index 4e4a851f67..901ceffbb2 100644
--- a/crates/ide-diagnostics/src/tests.rs
+++ b/crates/ide-diagnostics/src/tests.rs
@@ -198,12 +198,9 @@ pub(crate) fn check_diagnostics(ra_fixture: &str) {
}
#[track_caller]
-pub(crate) fn check_diagnostics_with_disabled(
- ra_fixture: &str,
- disabled: impl Iterator<Item = String>,
-) {
+pub(crate) fn check_diagnostics_with_disabled(ra_fixture: &str, disabled: &[&str]) {
let mut config = DiagnosticsConfig::test_sample();
- config.disabled.extend(disabled);
+ config.disabled.extend(disabled.iter().map(|&s| s.to_owned()));
check_diagnostics_with_config(config, ra_fixture)
}
diff --git a/crates/ide/src/hover.rs b/crates/ide/src/hover.rs
index 19b181ae3b..4a7350feb3 100644
--- a/crates/ide/src/hover.rs
+++ b/crates/ide/src/hover.rs
@@ -147,7 +147,7 @@ fn hover_simple(
if let Some(doc_comment) = token_as_doc_comment(&original_token) {
cov_mark::hit!(no_highlight_on_comment_hover);
return doc_comment.get_definition_with_descend_at(sema, offset, |def, node, range| {
- let res = hover_for_definition(sema, file_id, def, &node, config)?;
+ let res = hover_for_definition(sema, file_id, def, &node, config);
Some(RangeInfo::new(range, res))
});
}
@@ -161,7 +161,7 @@ fn hover_simple(
Definition::from(resolution?),
&original_token.parent()?,
config,
- )?;
+ );
return Some(RangeInfo::new(range, res));
}
@@ -215,7 +215,7 @@ fn hover_simple(
})
.flatten()
.unique_by(|&(def, _)| def)
- .filter_map(|(def, node)| hover_for_definition(sema, file_id, def, &node, config))
+ .map(|(def, node)| hover_for_definition(sema, file_id, def, &node, config))
.reduce(|mut acc: HoverResult, HoverResult { markup, actions }| {
acc.actions.extend(actions);
acc.markup = Markup::from(format!("{}\n---\n{markup}", acc.markup));
@@ -373,9 +373,9 @@ pub(crate) fn hover_for_definition(
def: Definition,
scope_node: &SyntaxNode,
config: &HoverConfig,
-) -> Option<HoverResult> {
+) -> HoverResult {
let famous_defs = match &def {
- Definition::BuiltinType(_) => Some(FamousDefs(sema, sema.scope(scope_node)?.krate())),
+ Definition::BuiltinType(_) => sema.scope(scope_node).map(|it| FamousDefs(sema, it.krate())),
_ => None,
};
@@ -396,20 +396,19 @@ pub(crate) fn hover_for_definition(
};
let notable_traits = def_ty.map(|ty| notable_traits(db, &ty)).unwrap_or_default();
- render::definition(sema.db, def, famous_defs.as_ref(), &notable_traits, config).map(|markup| {
- HoverResult {
- markup: render::process_markup(sema.db, def, &markup, config),
- actions: [
- show_implementations_action(sema.db, def),
- show_fn_references_action(sema.db, def),
- runnable_action(sema, def, file_id),
- goto_type_action_for_def(sema.db, def, &notable_traits),
- ]
- .into_iter()
- .flatten()
- .collect(),
- }
- })
+ let markup = render::definition(sema.db, def, famous_defs.as_ref(), &notable_traits, config);
+ HoverResult {
+ markup: render::process_markup(sema.db, def, &markup, config),
+ actions: [
+ show_implementations_action(sema.db, def),
+ show_fn_references_action(sema.db, def),
+ runnable_action(sema, def, file_id),
+ goto_type_action_for_def(sema.db, def, &notable_traits),
+ ]
+ .into_iter()
+ .flatten()
+ .collect(),
+ }
}
fn notable_traits(
diff --git a/crates/ide/src/hover/render.rs b/crates/ide/src/hover/render.rs
index eff055c959..563e78253a 100644
--- a/crates/ide/src/hover/render.rs
+++ b/crates/ide/src/hover/render.rs
@@ -3,8 +3,8 @@ use std::{mem, ops::Not};
use either::Either;
use hir::{
- Adt, AsAssocItem, CaptureKind, HasCrate, HasSource, HirDisplay, Layout, LayoutError, Name,
- Semantics, Trait, Type, TypeInfo,
+ Adt, AsAssocItem, AsExternAssocItem, CaptureKind, HasCrate, HasSource, HirDisplay, Layout,
+ LayoutError, Name, Semantics, Trait, Type, TypeInfo,
};
use ide_db::{
base_db::SourceDatabase,
@@ -264,7 +264,7 @@ pub(super) fn keyword(
let markup = process_markup(
sema.db,
Definition::Module(doc_owner),
- &markup(Some(docs.into()), description, None)?,
+ &markup(Some(docs.into()), description, None),
config,
);
Some(HoverResult { markup, actions })
@@ -369,12 +369,20 @@ fn definition_owner_name(db: &RootDatabase, def: &Definition) -> Option<String>
match def {
Definition::Field(f) => Some(f.parent_def(db).name(db)),
Definition::Local(l) => l.parent(db).name(db),
- Definition::Function(f) => match f.as_assoc_item(db)?.container(db) {
- hir::AssocItemContainer::Trait(t) => Some(t.name(db)),
- hir::AssocItemContainer::Impl(i) => i.self_ty(db).as_adt().map(|adt| adt.name(db)),
- },
Definition::Variant(e) => Some(e.parent_enum(db).name(db)),
- _ => None,
+
+ d => {
+ if let Some(assoc_item) = d.as_assoc_item(db) {
+ match assoc_item.container(db) {
+ hir::AssocItemContainer::Trait(t) => Some(t.name(db)),
+ hir::AssocItemContainer::Impl(i) => {
+ i.self_ty(db).as_adt().map(|adt| adt.name(db))
+ }
+ }
+ } else {
+ return d.as_extern_assoc_item(db).map(|_| "<extern>".to_owned());
+ }
+ }
}
.map(|name| name.display(db).to_string())
}
@@ -396,11 +404,11 @@ pub(super) fn definition(
famous_defs: Option<&FamousDefs<'_, '_>>,
notable_traits: &[(Trait, Vec<(Option<Type>, Name)>)],
config: &HoverConfig,
-) -> Option<Markup> {
+) -> Markup {
let mod_path = definition_mod_path(db, &def);
- let label = def.label(db)?;
+ let label = def.label(db);
let docs = def.docs(db, famous_defs);
- let value = match def {
+ let value = (|| match def {
Definition::Variant(it) => {
if !it.parent_enum(db).is_data_carrying(db) {
match it.eval(db) {
@@ -436,7 +444,7 @@ pub(super) fn definition(
Some(body.to_string())
}
_ => None,
- };
+ })();
let layout_info = match def {
Definition::Field(it) => render_memory_layout(
@@ -683,7 +691,7 @@ fn definition_mod_path(db: &RootDatabase, def: &Definition) -> Option<String> {
def.module(db).map(|module| path(db, module, definition_owner_name(db, def)))
}
-fn markup(docs: Option<String>, desc: String, mod_path: Option<String>) -> Option<Markup> {
+fn markup(docs: Option<String>, desc: String, mod_path: Option<String>) -> Markup {
let mut buf = String::new();
if let Some(mod_path) = mod_path {
@@ -696,7 +704,7 @@ fn markup(docs: Option<String>, desc: String, mod_path: Option<String>) -> Optio
if let Some(doc) = docs {
format_to!(buf, "\n___\n\n{}", doc);
}
- Some(buf.into())
+ buf.into()
}
fn find_std_module(famous_defs: &FamousDefs<'_, '_>, name: &str) -> Option<hir::Module> {
diff --git a/crates/ide/src/hover/tests.rs b/crates/ide/src/hover/tests.rs
index 69ddc1e45e..ead4f91595 100644
--- a/crates/ide/src/hover/tests.rs
+++ b/crates/ide/src/hover/tests.rs
@@ -1202,7 +1202,7 @@ fn main() {
*C*
```rust
- test
+ test::X
```
```rust
@@ -1279,11 +1279,11 @@ impl Thing {
);
check(
r#"
- enum Thing { A }
- impl Thing {
- pub fn thing(a: Self$0) {}
- }
- "#,
+enum Thing { A }
+impl Thing {
+ pub fn thing(a: Self$0) {}
+}
+"#,
expect![[r#"
*Self*
@@ -1298,6 +1298,42 @@ impl Thing {
```
"#]],
);
+ check(
+ r#"
+impl usize {
+ pub fn thing(a: Self$0) {}
+}
+"#,
+ expect![[r#"
+ *Self*
+
+ ```rust
+ test
+ ```
+
+ ```rust
+ usize
+ ```
+ "#]],
+ );
+ check(
+ r#"
+impl fn() -> usize {
+ pub fn thing(a: Self$0) {}
+}
+"#,
+ expect![[r#"
+ *Self*
+
+ ```rust
+ test
+ ```
+
+ ```rust
+ fn() -> usize
+ ```
+ "#]],
+ );
}
#[test]
@@ -2241,7 +2277,7 @@ fn main() { let foo_test = unsafe { fo$0o(1, 2, 3); } }
*foo*
```rust
- test
+ test::<extern>
```
```rust
@@ -4230,7 +4266,7 @@ fn main() {
*B*
```rust
- test
+ test::T
```
```rust
@@ -4259,7 +4295,7 @@ fn main() {
*B*
```rust
- test
+ test::T
```
```rust
@@ -4291,7 +4327,7 @@ fn main() {
*B*
```rust
- test
+ test::T
```
```rust
@@ -4883,7 +4919,7 @@ fn test() {
*FOO*
```rust
- test
+ test::S
```
```rust
@@ -5248,7 +5284,7 @@ impl T1 for Foo {
*Bar*
```rust
- test::t2
+ test::t2::T2
```
```rust
@@ -5270,7 +5306,7 @@ trait A {
*Assoc*
```rust
- test
+ test::A
```
```rust
@@ -5291,7 +5327,7 @@ trait A {
*Assoc*
```rust
- test
+ test::A
```
```rust
@@ -5310,7 +5346,7 @@ trait A where
*Assoc*
```rust
- test
+ test::A
```
```rust
@@ -6596,7 +6632,7 @@ fn test() {
*A*
```rust
- test
+ test::S
```
```rust
@@ -6625,7 +6661,7 @@ fn test() {
*A*
```rust
- test
+ test::S
```
```rust
@@ -6655,7 +6691,7 @@ mod m {
*A*
```rust
- test
+ test::S
```
```rust
@@ -7202,6 +7238,65 @@ impl Iterator for S {
}
#[test]
+fn extern_items() {
+ check(
+ r#"
+extern "C" {
+ static STATIC$0: ();
+}
+"#,
+ expect![[r#"
+ *STATIC*
+
+ ```rust
+ test::<extern>
+ ```
+
+ ```rust
+ static STATIC: ()
+ ```
+ "#]],
+ );
+ check(
+ r#"
+extern "C" {
+ fn fun$0();
+}
+"#,
+ expect![[r#"
+ *fun*
+
+ ```rust
+ test::<extern>
+ ```
+
+ ```rust
+ unsafe fn fun()
+ ```
+ "#]],
+ );
+ check(
+ r#"
+extern "C" {
+ type Ty$0;
+}
+"#,
+ expect![[r#"
+ *Ty*
+
+ ```rust
+ test::<extern>
+ ```
+
+ ```rust
+ // size = 0, align = 1
+ type Ty
+ ```
+ "#]],
+ );
+}
+
+#[test]
fn notable_ranged() {
check_hover_range(
r#"
diff --git a/crates/ide/src/join_lines.rs b/crates/ide/src/join_lines.rs
index fef0ec35ba..815a4ba7fd 100644
--- a/crates/ide/src/join_lines.rs
+++ b/crates/ide/src/join_lines.rs
@@ -303,7 +303,6 @@ fn compute_ws(left: SyntaxKind, right: SyntaxKind) -> &'static str {
#[cfg(test)]
mod tests {
- use syntax::SourceFile;
use test_utils::{add_cursor, assert_eq_text, extract_offset, extract_range};
use super::*;
diff --git a/crates/ide/src/rename.rs b/crates/ide/src/rename.rs
index f2eedfa431..f78153df38 100644
--- a/crates/ide/src/rename.rs
+++ b/crates/ide/src/rename.rs
@@ -9,6 +9,7 @@ use ide_db::{
base_db::{FileId, FileRange},
defs::{Definition, NameClass, NameRefClass},
rename::{bail, format_err, source_edit_from_references, IdentifierKind},
+ source_change::SourceChangeBuilder,
RootDatabase,
};
use itertools::Itertools;
@@ -90,24 +91,60 @@ pub(crate) fn rename(
let syntax = source_file.syntax();
let defs = find_definitions(&sema, syntax, position)?;
+ let alias_fallback = alias_fallback(syntax, position, new_name);
+
+ let ops: RenameResult<Vec<SourceChange>> = match alias_fallback {
+ Some(_) => defs
+ // FIXME: This can use the `ide_db::rename_reference` (or def.rename) method once we can
+ // properly find "direct" usages/references.
+ .map(|(.., def)| {
+ match IdentifierKind::classify(new_name)? {
+ IdentifierKind::Ident => (),
+ IdentifierKind::Lifetime => {
+ bail!("Cannot alias reference to a lifetime identifier")
+ }
+ IdentifierKind::Underscore => bail!("Cannot alias reference to `_`"),
+ };
- let ops: RenameResult<Vec<SourceChange>> = defs
- .map(|(.., def)| {
- if let Definition::Local(local) = def {
- if let Some(self_param) = local.as_self_param(sema.db) {
- cov_mark::hit!(rename_self_to_param);
- return rename_self_to_param(&sema, local, self_param, new_name);
- }
- if new_name == "self" {
- cov_mark::hit!(rename_to_self);
- return rename_to_self(&sema, local);
+ let mut usages = def.usages(&sema).all();
+
+ // FIXME: hack - removes the usage that triggered this rename operation.
+ match usages.references.get_mut(&position.file_id).and_then(|refs| {
+ refs.iter()
+ .position(|ref_| ref_.range.contains_inclusive(position.offset))
+ .map(|idx| refs.remove(idx))
+ }) {
+ Some(_) => (),
+ None => never!(),
+ };
+
+ let mut source_change = SourceChange::default();
+ source_change.extend(usages.iter().map(|(&file_id, refs)| {
+ (file_id, source_edit_from_references(refs, def, new_name))
+ }));
+
+ Ok(source_change)
+ })
+ .collect(),
+ None => defs
+ .map(|(.., def)| {
+ if let Definition::Local(local) = def {
+ if let Some(self_param) = local.as_self_param(sema.db) {
+ cov_mark::hit!(rename_self_to_param);
+ return rename_self_to_param(&sema, local, self_param, new_name);
+ }
+ if new_name == "self" {
+ cov_mark::hit!(rename_to_self);
+ return rename_to_self(&sema, local);
+ }
}
- }
- def.rename(&sema, new_name)
- })
- .collect();
+ def.rename(&sema, new_name)
+ })
+ .collect(),
+ };
ops?.into_iter()
+ .chain(alias_fallback)
.reduce(|acc, elem| acc.merge(elem))
.ok_or_else(|| format_err!("No references found at position"))
}
@@ -130,6 +167,38 @@ pub(crate) fn will_rename_file(
Some(change)
}
+// FIXME: Should support `extern crate`.
+fn alias_fallback(
+ syntax: &SyntaxNode,
+ FilePosition { file_id, offset }: FilePosition,
+ new_name: &str,
+) -> Option<SourceChange> {
+ let use_tree = syntax
+ .token_at_offset(offset)
+ .flat_map(|syntax| syntax.parent_ancestors())
+ .find_map(ast::UseTree::cast)?;
+
+ let last_path_segment = use_tree.path()?.segments().last()?.name_ref()?;
+ if !last_path_segment.syntax().text_range().contains_inclusive(offset) {
+ return None;
+ };
+
+ let mut builder = SourceChangeBuilder::new(file_id);
+
+ match use_tree.rename() {
+ Some(rename) => {
+ let offset = rename.syntax().text_range();
+ builder.replace(offset, format!("as {new_name}"));
+ }
+ None => {
+ let offset = use_tree.syntax().text_range().end();
+ builder.insert(offset, format!(" as {new_name}"));
+ }
+ }
+
+ Some(builder.finish())
+}
+
fn find_definitions(
sema: &Semantics<'_, RootDatabase>,
syntax: &SyntaxNode,
@@ -2626,7 +2695,8 @@ use qux as frob;
//- /lib.rs crate:lib new_source_root:library
pub struct S;
//- /main.rs crate:main deps:lib new_source_root:local
-use lib::S$0;
+use lib::S;
+fn main() { let _: S$0; }
"#,
"error: Cannot rename a non-local definition",
);
@@ -2686,4 +2756,27 @@ fn test() {
"#,
);
}
+
+ #[test]
+ fn rename_path_inside_use_tree() {
+ check(
+ "Baz",
+ r#"
+mod foo { pub struct Foo; }
+mod bar { use super::Foo; }
+
+use foo::Foo$0;
+
+fn main() { let _: Foo; }
+"#,
+ r#"
+mod foo { pub struct Foo; }
+mod bar { use super::Baz; }
+
+use foo::Foo as Baz;
+
+fn main() { let _: Baz; }
+"#,
+ )
+ }
}
diff --git a/crates/ide/src/static_index.rs b/crates/ide/src/static_index.rs
index 5feaf21aa9..2929a7522e 100644
--- a/crates/ide/src/static_index.rs
+++ b/crates/ide/src/static_index.rs
@@ -186,7 +186,7 @@ impl StaticIndex<'_> {
} else {
let it = self.tokens.insert(TokenStaticData {
documentation: documentation_for_definition(&sema, def, &node),
- hover: hover_for_definition(&sema, file_id, def, &node, &hover_config),
+ hover: Some(hover_for_definition(&sema, file_id, def, &node, &hover_config)),
definition: def.try_to_nav(self.db).map(UpmappingResult::call_site).map(|it| {
FileRange { file_id: it.file_id, range: it.focus_or_full_range() }
}),
@@ -196,7 +196,7 @@ impl StaticIndex<'_> {
enclosing_moniker: current_crate
.zip(def.enclosing_definition(self.db))
.and_then(|(cc, enclosing_def)| def_to_moniker(self.db, enclosing_def, cc)),
- signature: def.label(self.db),
+ signature: Some(def.label(self.db)),
kind: def_to_kind(self.db, def),
});
self.def_map.insert(def, it);
diff --git a/crates/load-cargo/src/lib.rs b/crates/load-cargo/src/lib.rs
index 8c5592da63..830d19a709 100644
--- a/crates/load-cargo/src/lib.rs
+++ b/crates/load-cargo/src/lib.rs
@@ -309,6 +309,10 @@ fn load_crate_graph(
vfs: &mut vfs::Vfs,
receiver: &Receiver<vfs::loader::Message>,
) -> AnalysisHost {
+ let (ProjectWorkspace::Cargo { toolchain, target_layout, .. }
+ | ProjectWorkspace::Json { toolchain, target_layout, .. }
+ | ProjectWorkspace::DetachedFiles { toolchain, target_layout, .. }) = ws;
+
let lru_cap = std::env::var("RA_LRU_CAP").ok().and_then(|it| it.parse::<usize>().ok());
let mut host = AnalysisHost::new(lru_cap);
let mut analysis_change = Change::new();
@@ -344,14 +348,9 @@ fn load_crate_graph(
let num_crates = crate_graph.len();
analysis_change.set_crate_graph(crate_graph);
analysis_change.set_proc_macros(proc_macros);
- if let ProjectWorkspace::Cargo { toolchain, target_layout, .. }
- | ProjectWorkspace::Json { toolchain, target_layout, .. } = ws
- {
- analysis_change.set_target_data_layouts(
- iter::repeat(target_layout.clone()).take(num_crates).collect(),
- );
- analysis_change.set_toolchains(iter::repeat(toolchain.clone()).take(num_crates).collect());
- }
+ analysis_change
+ .set_target_data_layouts(iter::repeat(target_layout.clone()).take(num_crates).collect());
+ analysis_change.set_toolchains(iter::repeat(toolchain.clone()).take(num_crates).collect());
host.apply_change(analysis_change);
host
diff --git a/crates/proc-macro-api/src/process.rs b/crates/proc-macro-api/src/process.rs
index 12eafcea44..72f95643c8 100644
--- a/crates/proc-macro-api/src/process.rs
+++ b/crates/proc-macro-api/src/process.rs
@@ -45,7 +45,7 @@ impl ProcMacroProcessSrv {
})
};
let mut srv = create_srv(true)?;
- tracing::info!("sending version check");
+ tracing::info!("sending proc-macro server version check");
match srv.version_check() {
Ok(v) if v > CURRENT_API_VERSION => Err(io::Error::new(
io::ErrorKind::Other,
@@ -55,14 +55,15 @@ impl ProcMacroProcessSrv {
),
)),
Ok(v) => {
- tracing::info!("got version {v}");
+ tracing::info!("Proc-macro server version: {v}");
srv = create_srv(false)?;
srv.version = v;
- if srv.version > RUST_ANALYZER_SPAN_SUPPORT {
+ if srv.version >= RUST_ANALYZER_SPAN_SUPPORT {
if let Ok(mode) = srv.enable_rust_analyzer_spans() {
srv.mode = mode;
}
}
+ tracing::info!("Proc-macro server span mode: {:?}", srv.mode);
Ok(srv)
}
Err(e) => {
diff --git a/crates/proc-macro-srv/src/proc_macros.rs b/crates/proc-macro-srv/src/proc_macros.rs
index 3fe968c81c..686d5b0438 100644
--- a/crates/proc-macro-srv/src/proc_macros.rs
+++ b/crates/proc-macro-srv/src/proc_macros.rs
@@ -64,7 +64,7 @@ impl ProcMacros {
&bridge::server::SameThread,
S::make_server(call_site, def_site, mixed_site),
parsed_body,
- false,
+ cfg!(debug_assertions),
);
return res
.map(|it| it.into_subtree(call_site))
@@ -75,7 +75,7 @@ impl ProcMacros {
&bridge::server::SameThread,
S::make_server(call_site, def_site, mixed_site),
parsed_body,
- false,
+ cfg!(debug_assertions),
);
return res
.map(|it| it.into_subtree(call_site))
@@ -87,7 +87,7 @@ impl ProcMacros {
S::make_server(call_site, def_site, mixed_site),
parsed_attributes,
parsed_body,
- false,
+ cfg!(debug_assertions),
);
return res
.map(|it| it.into_subtree(call_site))
diff --git a/crates/proc-macro-srv/src/server.rs b/crates/proc-macro-srv/src/server.rs
index ff8fd295d8..5a814e23e7 100644
--- a/crates/proc-macro-srv/src/server.rs
+++ b/crates/proc-macro-srv/src/server.rs
@@ -93,7 +93,14 @@ impl<S> LiteralFormatter<S> {
let hashes = get_hashes_str(n);
f(&["br", hashes, "\"", symbol, "\"", hashes, suffix])
}
- _ => f(&[symbol, suffix]),
+ bridge::LitKind::CStr => f(&["c\"", symbol, "\"", suffix]),
+ bridge::LitKind::CStrRaw(n) => {
+ let hashes = get_hashes_str(n);
+ f(&["cr", hashes, "\"", symbol, "\"", hashes, suffix])
+ }
+ bridge::LitKind::Integer | bridge::LitKind::Float | bridge::LitKind::ErrWithGuar => {
+ f(&[symbol, suffix])
+ }
})
}
diff --git a/crates/proc-macro-srv/src/server/rust_analyzer_span.rs b/crates/proc-macro-srv/src/server/rust_analyzer_span.rs
index c6a0a66655..15d260d518 100644
--- a/crates/proc-macro-srv/src/server/rust_analyzer_span.rs
+++ b/crates/proc-macro-srv/src/server/rust_analyzer_span.rs
@@ -10,16 +10,16 @@ use std::{
ops::{Bound, Range},
};
-use ::tt::{TextRange, TextSize};
use proc_macro::bridge::{self, server};
use span::{Span, FIXUP_ERASED_FILE_AST_ID_MARKER};
+use tt::{TextRange, TextSize};
use crate::server::{
delim_to_external, delim_to_internal, token_stream::TokenStreamBuilder, LiteralFormatter,
Symbol, SymbolInternerRef, SYMBOL_INTERNER,
};
mod tt {
- pub use ::tt::*;
+ pub use tt::*;
pub type Subtree = ::tt::Subtree<super::Span>;
pub type TokenTree = ::tt::TokenTree<super::Span>;
@@ -97,22 +97,33 @@ impl server::FreeFunctions for RaSpanServer {
}
let TokenKind::Literal { kind, suffix_start } = lit.kind else { return Err(()) };
- let kind = match kind {
- LiteralKind::Int { .. } => LitKind::Integer,
- LiteralKind::Float { .. } => LitKind::Float,
- LiteralKind::Char { .. } => LitKind::Char,
- LiteralKind::Byte { .. } => LitKind::Byte,
- LiteralKind::Str { .. } => LitKind::Str,
- LiteralKind::ByteStr { .. } => LitKind::ByteStr,
- LiteralKind::CStr { .. } => LitKind::CStr,
- LiteralKind::RawStr { n_hashes } => LitKind::StrRaw(n_hashes.unwrap_or_default()),
- LiteralKind::RawByteStr { n_hashes } => {
- LitKind::ByteStrRaw(n_hashes.unwrap_or_default())
- }
- LiteralKind::RawCStr { n_hashes } => LitKind::CStrRaw(n_hashes.unwrap_or_default()),
+ let (kind, start_offset, end_offset) = match kind {
+ LiteralKind::Int { .. } => (LitKind::Integer, 0, 0),
+ LiteralKind::Float { .. } => (LitKind::Float, 0, 0),
+ LiteralKind::Char { terminated } => (LitKind::Char, 1, terminated as usize),
+ LiteralKind::Byte { terminated } => (LitKind::Byte, 2, terminated as usize),
+ LiteralKind::Str { terminated } => (LitKind::Str, 1, terminated as usize),
+ LiteralKind::ByteStr { terminated } => (LitKind::ByteStr, 2, terminated as usize),
+ LiteralKind::CStr { terminated } => (LitKind::CStr, 2, terminated as usize),
+ LiteralKind::RawStr { n_hashes } => (
+ LitKind::StrRaw(n_hashes.unwrap_or_default()),
+ 2 + n_hashes.unwrap_or_default() as usize,
+ 1 + n_hashes.unwrap_or_default() as usize,
+ ),
+ LiteralKind::RawByteStr { n_hashes } => (
+ LitKind::ByteStrRaw(n_hashes.unwrap_or_default()),
+ 3 + n_hashes.unwrap_or_default() as usize,
+ 1 + n_hashes.unwrap_or_default() as usize,
+ ),
+ LiteralKind::RawCStr { n_hashes } => (
+ LitKind::CStrRaw(n_hashes.unwrap_or_default()),
+ 3 + n_hashes.unwrap_or_default() as usize,
+ 1 + n_hashes.unwrap_or_default() as usize,
+ ),
};
let (lit, suffix) = s.split_at(suffix_start as usize);
+ let lit = &lit[start_offset..lit.len() - end_offset];
let suffix = match suffix {
"" | "_" => None,
suffix => Some(Symbol::intern(self.interner, suffix)),
@@ -248,12 +259,8 @@ impl server::TokenStream for RaSpanServer {
}
tt::TokenTree::Leaf(tt::Leaf::Literal(lit)) => {
bridge::TokenTree::Literal(bridge::Literal {
- // FIXME: handle literal kinds
- kind: bridge::LitKind::Integer, // dummy
- symbol: Symbol::intern(self.interner, &lit.text),
- // FIXME: handle suffixes
- suffix: None,
span: lit.span,
+ ..server::FreeFunctions::literal_from_str(self, &lit.text).unwrap()
})
}
tt::TokenTree::Leaf(tt::Leaf::Punct(punct)) => {
diff --git a/crates/proc-macro-srv/src/server/token_id.rs b/crates/proc-macro-srv/src/server/token_id.rs
index 7e9d8057ac..f40c850b25 100644
--- a/crates/proc-macro-srv/src/server/token_id.rs
+++ b/crates/proc-macro-srv/src/server/token_id.rs
@@ -14,7 +14,7 @@ use crate::server::{
mod tt {
pub use proc_macro_api::msg::TokenId;
- pub use ::tt::*;
+ pub use tt::*;
pub type Subtree = ::tt::Subtree<TokenId>;
pub type TokenTree = ::tt::TokenTree<TokenId>;
@@ -89,22 +89,34 @@ impl server::FreeFunctions for TokenIdServer {
}
let TokenKind::Literal { kind, suffix_start } = lit.kind else { return Err(()) };
- let kind = match kind {
- LiteralKind::Int { .. } => LitKind::Integer,
- LiteralKind::Float { .. } => LitKind::Float,
- LiteralKind::Char { .. } => LitKind::Char,
- LiteralKind::Byte { .. } => LitKind::Byte,
- LiteralKind::Str { .. } => LitKind::Str,
- LiteralKind::ByteStr { .. } => LitKind::ByteStr,
- LiteralKind::CStr { .. } => LitKind::CStr,
- LiteralKind::RawStr { n_hashes } => LitKind::StrRaw(n_hashes.unwrap_or_default()),
- LiteralKind::RawByteStr { n_hashes } => {
- LitKind::ByteStrRaw(n_hashes.unwrap_or_default())
- }
- LiteralKind::RawCStr { n_hashes } => LitKind::CStrRaw(n_hashes.unwrap_or_default()),
+
+ let (kind, start_offset, end_offset) = match kind {
+ LiteralKind::Int { .. } => (LitKind::Integer, 0, 0),
+ LiteralKind::Float { .. } => (LitKind::Float, 0, 0),
+ LiteralKind::Char { terminated } => (LitKind::Char, 1, terminated as usize),
+ LiteralKind::Byte { terminated } => (LitKind::Byte, 2, terminated as usize),
+ LiteralKind::Str { terminated } => (LitKind::Str, 1, terminated as usize),
+ LiteralKind::ByteStr { terminated } => (LitKind::ByteStr, 2, terminated as usize),
+ LiteralKind::CStr { terminated } => (LitKind::CStr, 2, terminated as usize),
+ LiteralKind::RawStr { n_hashes } => (
+ LitKind::StrRaw(n_hashes.unwrap_or_default()),
+ 2 + n_hashes.unwrap_or_default() as usize,
+ 1 + n_hashes.unwrap_or_default() as usize,
+ ),
+ LiteralKind::RawByteStr { n_hashes } => (
+ LitKind::ByteStrRaw(n_hashes.unwrap_or_default()),
+ 3 + n_hashes.unwrap_or_default() as usize,
+ 1 + n_hashes.unwrap_or_default() as usize,
+ ),
+ LiteralKind::RawCStr { n_hashes } => (
+ LitKind::CStrRaw(n_hashes.unwrap_or_default()),
+ 3 + n_hashes.unwrap_or_default() as usize,
+ 1 + n_hashes.unwrap_or_default() as usize,
+ ),
};
let (lit, suffix) = s.split_at(suffix_start as usize);
+ let lit = &lit[start_offset..lit.len() - end_offset];
let suffix = match suffix {
"" | "_" => None,
suffix => Some(Symbol::intern(self.interner, suffix)),
@@ -233,12 +245,9 @@ impl server::TokenStream for TokenIdServer {
}
tt::TokenTree::Leaf(tt::Leaf::Literal(lit)) => {
bridge::TokenTree::Literal(bridge::Literal {
- // FIXME: handle literal kinds
- kind: bridge::LitKind::Integer, // dummy
- symbol: Symbol::intern(self.interner, &lit.text),
- // FIXME: handle suffixes
- suffix: None,
span: lit.span,
+ ..server::FreeFunctions::literal_from_str(self, &lit.text)
+ .unwrap_or_else(|_| panic!("`{}`", lit.text))
})
}
tt::TokenTree::Leaf(tt::Leaf::Punct(punct)) => {
diff --git a/crates/proc-macro-srv/src/server/token_stream.rs b/crates/proc-macro-srv/src/server/token_stream.rs
index 5edaa720fc..408db60e87 100644
--- a/crates/proc-macro-srv/src/server/token_stream.rs
+++ b/crates/proc-macro-srv/src/server/token_stream.rs
@@ -115,8 +115,6 @@ pub(super) mod token_stream {
}
}
- type LexError = String;
-
/// Attempts to break the string into tokens and parse those tokens into a token stream.
/// May fail for a number of reasons, for example, if the string contains unbalanced delimiters
/// or characters not existing in the language.
@@ -124,13 +122,10 @@ pub(super) mod token_stream {
///
/// NOTE: some errors may cause panics instead of returning `LexError`. We reserve the right to
/// change these errors into `LexError`s later.
- #[rustfmt::skip]
- impl<S: tt::Span> /*FromStr for*/ TokenStream<S> {
- // type Err = LexError;
-
- pub(crate) fn from_str(src: &str, call_site: S) -> Result<TokenStream<S>, LexError> {
+ impl<S: tt::Span> TokenStream<S> {
+ pub(crate) fn from_str(src: &str, call_site: S) -> Result<TokenStream<S>, String> {
let subtree =
- mbe::parse_to_token_tree_static_span(call_site, src).ok_or("Failed to parse from mbe")?;
+ mbe::parse_to_token_tree_static_span(call_site, src).ok_or("lexing error")?;
Ok(TokenStream::with_subtree(subtree))
}
diff --git a/crates/proc-macro-srv/src/tests/mod.rs b/crates/proc-macro-srv/src/tests/mod.rs
index e5bfe5ee92..54a20357d2 100644
--- a/crates/proc-macro-srv/src/tests/mod.rs
+++ b/crates/proc-macro-srv/src/tests/mod.rs
@@ -169,7 +169,7 @@ fn test_fn_like_mk_idents() {
fn test_fn_like_macro_clone_literals() {
assert_expand(
"fn_like_clone_tokens",
- r###"1u16, 2_u32, -4i64, 3.14f32, "hello bridge", "suffixed"suffix, r##"raw"##"###,
+ r###"1u16, 2_u32, -4i64, 3.14f32, "hello bridge", "suffixed"suffix, r##"raw"##, 'a', b'b', c"null""###,
expect![[r###"
SUBTREE $$ 1 1
LITERAL 1u16 1
@@ -181,11 +181,17 @@ fn test_fn_like_macro_clone_literals() {
PUNCH , [alone] 1
LITERAL 3.14f32 1
PUNCH , [alone] 1
- LITERAL ""hello bridge"" 1
+ LITERAL "hello bridge" 1
PUNCH , [alone] 1
- LITERAL ""suffixed""suffix 1
+ LITERAL "suffixed"suffix 1
PUNCH , [alone] 1
- LITERAL r##"r##"raw"##"## 1"###]],
+ LITERAL r##"raw"## 1
+ PUNCH , [alone] 1
+ LITERAL 'a' 1
+ PUNCH , [alone] 1
+ LITERAL b'b' 1
+ PUNCH , [alone] 1
+ LITERAL c"null" 1"###]],
expect![[r###"
SUBTREE $$ SpanData { range: 0..100, anchor: SpanAnchor(FileId(42), 2), ctx: SyntaxContextId(0) } SpanData { range: 0..100, anchor: SpanAnchor(FileId(42), 2), ctx: SyntaxContextId(0) }
LITERAL 1u16 SpanData { range: 0..4, anchor: SpanAnchor(FileId(42), 2), ctx: SyntaxContextId(0) }
@@ -197,11 +203,17 @@ fn test_fn_like_macro_clone_literals() {
PUNCH , [alone] SpanData { range: 18..19, anchor: SpanAnchor(FileId(42), 2), ctx: SyntaxContextId(0) }
LITERAL 3.14f32 SpanData { range: 20..27, anchor: SpanAnchor(FileId(42), 2), ctx: SyntaxContextId(0) }
PUNCH , [alone] SpanData { range: 27..28, anchor: SpanAnchor(FileId(42), 2), ctx: SyntaxContextId(0) }
- LITERAL ""hello bridge"" SpanData { range: 29..43, anchor: SpanAnchor(FileId(42), 2), ctx: SyntaxContextId(0) }
+ LITERAL "hello bridge" SpanData { range: 29..43, anchor: SpanAnchor(FileId(42), 2), ctx: SyntaxContextId(0) }
PUNCH , [alone] SpanData { range: 43..44, anchor: SpanAnchor(FileId(42), 2), ctx: SyntaxContextId(0) }
- LITERAL ""suffixed""suffix SpanData { range: 45..61, anchor: SpanAnchor(FileId(42), 2), ctx: SyntaxContextId(0) }
+ LITERAL "suffixed"suffix SpanData { range: 45..61, anchor: SpanAnchor(FileId(42), 2), ctx: SyntaxContextId(0) }
PUNCH , [alone] SpanData { range: 61..62, anchor: SpanAnchor(FileId(42), 2), ctx: SyntaxContextId(0) }
- LITERAL r##"r##"raw"##"## SpanData { range: 63..73, anchor: SpanAnchor(FileId(42), 2), ctx: SyntaxContextId(0) }"###]],
+ LITERAL r##"raw"## SpanData { range: 63..73, anchor: SpanAnchor(FileId(42), 2), ctx: SyntaxContextId(0) }
+ PUNCH , [alone] SpanData { range: 73..74, anchor: SpanAnchor(FileId(42), 2), ctx: SyntaxContextId(0) }
+ LITERAL 'a' SpanData { range: 75..78, anchor: SpanAnchor(FileId(42), 2), ctx: SyntaxContextId(0) }
+ PUNCH , [alone] SpanData { range: 78..79, anchor: SpanAnchor(FileId(42), 2), ctx: SyntaxContextId(0) }
+ LITERAL b'b' SpanData { range: 80..84, anchor: SpanAnchor(FileId(42), 2), ctx: SyntaxContextId(0) }
+ PUNCH , [alone] SpanData { range: 84..85, anchor: SpanAnchor(FileId(42), 2), ctx: SyntaxContextId(0) }
+ LITERAL c"null" SpanData { range: 86..93, anchor: SpanAnchor(FileId(42), 2), ctx: SyntaxContextId(0) }"###]],
);
}
diff --git a/crates/project-model/src/build_scripts.rs b/crates/project-model/src/build_scripts.rs
index ab72f1fba0..621b6ca3ef 100644
--- a/crates/project-model/src/build_scripts.rs
+++ b/crates/project-model/src/build_scripts.rs
@@ -138,7 +138,7 @@ impl WorkspaceBuildScripts {
toolchain: &Option<Version>,
sysroot: Option<&Sysroot>,
) -> io::Result<WorkspaceBuildScripts> {
- const RUST_1_62: Version = Version::new(1, 62, 0);
+ const RUST_1_75: Version = Version::new(1, 75, 0);
let current_dir = match &config.invocation_location {
InvocationLocation::Root(root) if config.run_build_script_command.is_some() => {
@@ -162,7 +162,7 @@ impl WorkspaceBuildScripts {
progress,
) {
Ok(WorkspaceBuildScripts { error: Some(error), .. })
- if toolchain.as_ref().map_or(false, |it| *it >= RUST_1_62) =>
+ if toolchain.as_ref().map_or(false, |it| *it >= RUST_1_75) =>
{
// building build scripts failed, attempt to build with --keep-going so
// that we potentially get more build data
@@ -172,7 +172,8 @@ impl WorkspaceBuildScripts {
&workspace.workspace_root().to_path_buf(),
sysroot,
)?;
- cmd.args(["-Z", "unstable-options", "--keep-going"]).env("RUSTC_BOOTSTRAP", "1");
+
+ cmd.args(["--keep-going"]);
let mut res = Self::run_per_ws(cmd, workspace, current_dir, progress)?;
res.error = Some(error);
Ok(res)
diff --git a/crates/project-model/src/target_data_layout.rs b/crates/project-model/src/target_data_layout.rs
index af635dda57..98917351c5 100644
--- a/crates/project-model/src/target_data_layout.rs
+++ b/crates/project-model/src/target_data_layout.rs
@@ -32,7 +32,16 @@ pub fn get(
Sysroot::set_rustup_toolchain_env(&mut cmd, sysroot);
cmd.envs(extra_env);
cmd.current_dir(cargo_toml.parent())
- .args(["rustc", "--", "-Z", "unstable-options", "--print", "target-spec-json"])
+ .args([
+ "rustc",
+ "-Z",
+ "unstable-options",
+ "--print",
+ "target-spec-json",
+ "--",
+ "-Z",
+ "unstable-options",
+ ])
.env("RUSTC_BOOTSTRAP", "1");
if let Some(target) = target {
cmd.args(["--target", target]);
diff --git a/crates/project-model/src/workspace.rs b/crates/project-model/src/workspace.rs
index b7ae76be8c..bcb5dcadb5 100644
--- a/crates/project-model/src/workspace.rs
+++ b/crates/project-model/src/workspace.rs
@@ -100,6 +100,8 @@ pub enum ProjectWorkspace {
/// Holds cfg flags for the current target. We get those by running
/// `rustc --print cfg`.
rustc_cfg: Vec<CfgFlag>,
+ toolchain: Option<Version>,
+ target_layout: TargetLayoutLoadResult,
},
}
@@ -145,16 +147,24 @@ impl fmt::Debug for ProjectWorkspace {
debug_struct.field("n_sysroot_crates", &sysroot.num_packages());
}
debug_struct
- .field("toolchain", &toolchain)
.field("n_rustc_cfg", &rustc_cfg.len())
+ .field("toolchain", &toolchain)
.field("data_layout", &data_layout);
debug_struct.finish()
}
- ProjectWorkspace::DetachedFiles { files, sysroot, rustc_cfg } => f
+ ProjectWorkspace::DetachedFiles {
+ files,
+ sysroot,
+ rustc_cfg,
+ toolchain,
+ target_layout,
+ } => f
.debug_struct("DetachedFiles")
.field("n_files", &files.len())
.field("sysroot", &sysroot.is_ok())
.field("n_rustc_cfg", &rustc_cfg.len())
+ .field("toolchain", &toolchain)
+ .field("data_layout", &target_layout)
.finish(),
}
}
@@ -403,32 +413,54 @@ impl ProjectWorkspace {
detached_files: Vec<AbsPathBuf>,
config: &CargoConfig,
) -> anyhow::Result<ProjectWorkspace> {
+ let dir = detached_files
+ .first()
+ .and_then(|it| it.parent())
+ .ok_or_else(|| format_err!("No detached files to load"))?;
let sysroot = match &config.sysroot {
Some(RustLibSource::Path(path)) => {
Sysroot::with_sysroot_dir(path.clone(), config.sysroot_query_metadata)
.map_err(|e| Some(format!("Failed to find sysroot at {path}:{e}")))
}
- Some(RustLibSource::Discover) => {
- let dir = &detached_files
- .first()
- .and_then(|it| it.parent())
- .ok_or_else(|| format_err!("No detached files to load"))?;
- Sysroot::discover(dir, &config.extra_env, config.sysroot_query_metadata).map_err(
- |e| {
- Some(format!(
- "Failed to find sysroot for {dir}. Is rust-src installed? {e}"
- ))
- },
- )
- }
+ Some(RustLibSource::Discover) => Sysroot::discover(
+ dir,
+ &config.extra_env,
+ config.sysroot_query_metadata,
+ )
+ .map_err(|e| {
+ Some(format!("Failed to find sysroot for {dir}. Is rust-src installed? {e}"))
+ }),
None => Err(None),
};
- let rustc_cfg = rustc_cfg::get(
+
+ let sysroot_ref = sysroot.as_ref().ok();
+ let toolchain = match get_toolchain_version(
+ dir,
+ sysroot_ref,
+ toolchain::Tool::Rustc,
+ &config.extra_env,
+ "rustc ",
+ ) {
+ Ok(it) => it,
+ Err(e) => {
+ tracing::error!("{e}");
+ None
+ }
+ };
+
+ let rustc_cfg = rustc_cfg::get(None, &config.extra_env, RustcCfgConfig::Rustc(sysroot_ref));
+ let data_layout = target_data_layout::get(
+ RustcDataLayoutConfig::Rustc(sysroot_ref),
None,
- &FxHashMap::default(),
- RustcCfgConfig::Rustc(sysroot.as_ref().ok()),
+ &config.extra_env,
);
- Ok(ProjectWorkspace::DetachedFiles { files: detached_files, sysroot, rustc_cfg })
+ Ok(ProjectWorkspace::DetachedFiles {
+ files: detached_files,
+ sysroot,
+ rustc_cfg,
+ toolchain,
+ target_layout: data_layout.map(Arc::from).map_err(|it| Arc::from(it.to_string())),
+ })
}
/// Runs the build scripts for this [`ProjectWorkspace`].
@@ -724,7 +756,13 @@ impl ProjectWorkspace {
cfg_overrides,
build_scripts,
),
- ProjectWorkspace::DetachedFiles { files, sysroot, rustc_cfg } => {
+ ProjectWorkspace::DetachedFiles {
+ files,
+ sysroot,
+ rustc_cfg,
+ toolchain: _,
+ target_layout: _,
+ } => {
detached_files_to_crate_graph(rustc_cfg.clone(), load, files, sysroot.as_ref().ok())
}
};
@@ -786,9 +824,21 @@ impl ProjectWorkspace {
&& toolchain == o_toolchain
}
(
- Self::DetachedFiles { files, sysroot, rustc_cfg },
- Self::DetachedFiles { files: o_files, sysroot: o_sysroot, rustc_cfg: o_rustc_cfg },
- ) => files == o_files && sysroot == o_sysroot && rustc_cfg == o_rustc_cfg,
+ Self::DetachedFiles { files, sysroot, rustc_cfg, toolchain, target_layout },
+ Self::DetachedFiles {
+ files: o_files,
+ sysroot: o_sysroot,
+ rustc_cfg: o_rustc_cfg,
+ toolchain: o_toolchain,
+ target_layout: o_target_layout,
+ },
+ ) => {
+ files == o_files
+ && sysroot == o_sysroot
+ && rustc_cfg == o_rustc_cfg
+ && toolchain == o_toolchain
+ && target_layout == o_target_layout
+ }
_ => false,
}
}
diff --git a/crates/rust-analyzer/src/cargo_target_spec.rs b/crates/rust-analyzer/src/cargo_target_spec.rs
index f9f2617825..815a98980b 100644
--- a/crates/rust-analyzer/src/cargo_target_spec.rs
+++ b/crates/rust-analyzer/src/cargo_target_spec.rs
@@ -208,7 +208,6 @@ fn required_features(cfg_expr: &CfgExpr, features: &mut Vec<String>) {
mod tests {
use super::*;
- use cfg::CfgExpr;
use mbe::{syntax_node_to_token_tree, DummyTestSpanMap, DUMMY};
use syntax::{
ast::{self, AstNode},
diff --git a/crates/rust-analyzer/src/cli/flags.rs b/crates/rust-analyzer/src/cli/flags.rs
index 493e614dce..3f68c5d053 100644
--- a/crates/rust-analyzer/src/cli/flags.rs
+++ b/crates/rust-analyzer/src/cli/flags.rs
@@ -30,7 +30,7 @@ xflags::xflags! {
default cmd lsp-server {
/// Print version.
- optional --version
+ optional -V, --version
/// Dump a LSP config JSON schema.
optional --print-config-schema
diff --git a/crates/rust-analyzer/src/cli/rustc_tests.rs b/crates/rust-analyzer/src/cli/rustc_tests.rs
index 64ea246a45..7062b60cbf 100644
--- a/crates/rust-analyzer/src/cli/rustc_tests.rs
+++ b/crates/rust-analyzer/src/cli/rustc_tests.rs
@@ -1,11 +1,16 @@
//! Run all tests in a project, similar to `cargo test`, but using the mir interpreter.
+use std::convert::identity;
+use std::thread::Builder;
+use std::time::{Duration, Instant};
use std::{cell::RefCell, fs::read_to_string, panic::AssertUnwindSafe, path::PathBuf};
use hir::{Change, Crate};
use ide::{AnalysisHost, DiagnosticCode, DiagnosticsConfig};
+use itertools::Either;
use profile::StopWatch;
-use project_model::{CargoConfig, ProjectWorkspace, RustLibSource, Sysroot};
+use project_model::target_data_layout::RustcDataLayoutConfig;
+use project_model::{target_data_layout, CargoConfig, ProjectWorkspace, RustLibSource, Sysroot};
use load_cargo::{load_workspace, LoadCargoConfig, ProcMacroServerChoice};
use rustc_hash::FxHashMap;
@@ -60,15 +65,22 @@ impl Tester {
std::fs::write(&tmp_file, "")?;
let cargo_config =
CargoConfig { sysroot: Some(RustLibSource::Discover), ..Default::default() };
+
+ let sysroot =
+ Ok(Sysroot::discover(tmp_file.parent().unwrap(), &cargo_config.extra_env, false)
+ .unwrap());
+ let data_layout = target_data_layout::get(
+ RustcDataLayoutConfig::Rustc(sysroot.as_ref().ok()),
+ None,
+ &cargo_config.extra_env,
+ );
+
let workspace = ProjectWorkspace::DetachedFiles {
files: vec![tmp_file.clone()],
- sysroot: Ok(Sysroot::discover(
- tmp_file.parent().unwrap(),
- &cargo_config.extra_env,
- false,
- )
- .unwrap()),
+ sysroot,
rustc_cfg: vec![],
+ toolchain: None,
+ target_layout: data_layout.map(Arc::from).map_err(|it| Arc::from(it.to_string())),
};
let load_cargo_config = LoadCargoConfig {
load_out_dirs_from_check: false,
@@ -92,6 +104,7 @@ impl Tester {
}
fn test(&mut self, p: PathBuf) {
+ println!("{}", p.display());
if p.parent().unwrap().file_name().unwrap() == "auxiliary" {
// These are not tests
return;
@@ -124,15 +137,44 @@ impl Tester {
self.host.apply_change(change);
let diagnostic_config = DiagnosticsConfig::test_sample();
+ let res = std::thread::scope(|s| {
+ let worker = Builder::new()
+ .stack_size(40 * 1024 * 1024)
+ .spawn_scoped(s, {
+ let diagnostic_config = &diagnostic_config;
+ let main = std::thread::current();
+ let analysis = self.host.analysis();
+ let root_file = self.root_file;
+ move || {
+ let res = std::panic::catch_unwind(move || {
+ analysis.diagnostics(
+ diagnostic_config,
+ ide::AssistResolveStrategy::None,
+ root_file,
+ )
+ });
+ main.unpark();
+ res
+ }
+ })
+ .unwrap();
+
+ let timeout = Duration::from_secs(5);
+ let now = Instant::now();
+ while now.elapsed() <= timeout && !worker.is_finished() {
+ std::thread::park_timeout(timeout - now.elapsed());
+ }
+
+ if !worker.is_finished() {
+ // attempt to cancel the worker, won't work for chalk hangs unfortunately
+ self.host.request_cancellation();
+ }
+ worker.join().and_then(identity)
+ });
let mut actual = FxHashMap::default();
- let panicked = match std::panic::catch_unwind(|| {
- self.host
- .analysis()
- .diagnostics(&diagnostic_config, ide::AssistResolveStrategy::None, self.root_file)
- .unwrap()
- }) {
- Err(e) => Some(e),
- Ok(diags) => {
+ let panicked = match res {
+ Err(e) => Some(Either::Left(e)),
+ Ok(Ok(diags)) => {
for diag in diags {
if !matches!(diag.code, DiagnosticCode::RustcHardError(_)) {
continue;
@@ -144,6 +186,7 @@ impl Tester {
}
None
}
+ Ok(Err(e)) => Some(Either::Right(e)),
};
// Ignore tests with diagnostics that we don't emit.
ignore_test |= expected.keys().any(|k| !SUPPORTED_DIAGNOSTICS.contains(k));
@@ -151,14 +194,19 @@ impl Tester {
println!("{p:?} IGNORE");
self.ignore_count += 1;
} else if let Some(panic) = panicked {
- if let Some(msg) = panic
- .downcast_ref::<String>()
- .map(String::as_str)
- .or_else(|| panic.downcast_ref::<&str>().copied())
- {
- println!("{msg:?} ")
+ match panic {
+ Either::Left(panic) => {
+ if let Some(msg) = panic
+ .downcast_ref::<String>()
+ .map(String::as_str)
+ .or_else(|| panic.downcast_ref::<&str>().copied())
+ {
+ println!("{msg:?} ")
+ }
+ println!("{p:?} PANIC");
+ }
+ Either::Right(_) => println!("{p:?} CANCELLED"),
}
- println!("PANIC");
self.fail_count += 1;
} else if actual == expected {
println!("{p:?} PASS");
@@ -228,6 +276,7 @@ impl flags::RustcTests {
pub fn run(self) -> Result<()> {
let mut tester = Tester::new()?;
let walk_dir = WalkDir::new(self.rustc_repo.join("tests/ui"));
+ eprintln!("Running tests for tests/ui");
for i in walk_dir {
let i = i?;
let p = i.into_path();
diff --git a/crates/rust-analyzer/src/cli/scip.rs b/crates/rust-analyzer/src/cli/scip.rs
index 2d56830c87..27869a5a7e 100644
--- a/crates/rust-analyzer/src/cli/scip.rs
+++ b/crates/rust-analyzer/src/cli/scip.rs
@@ -324,7 +324,7 @@ fn moniker_to_symbol(moniker: &MonikerResult) -> scip_types::Symbol {
#[cfg(test)]
mod test {
use super::*;
- use ide::{AnalysisHost, FilePosition, StaticIndex, TextSize};
+ use ide::{AnalysisHost, FilePosition, TextSize};
use scip::symbol::format_symbol;
use test_fixture::ChangeFixture;
diff --git a/crates/rust-analyzer/src/global_state.rs b/crates/rust-analyzer/src/global_state.rs
index 293807a383..b2d507491b 100644
--- a/crates/rust-analyzer/src/global_state.rs
+++ b/crates/rust-analyzer/src/global_state.rs
@@ -301,19 +301,12 @@ impl GlobalState {
if let Some(path) = vfs_path.as_path() {
let path = path.to_path_buf();
if reload::should_refresh_for_change(&path, file.kind()) {
- workspace_structure_change = Some((
- path.clone(),
- false,
- AsRef::<std::path::Path>::as_ref(&path).ends_with("build.rs"),
- ));
+ workspace_structure_change = Some((path.clone(), false));
}
if file.is_created_or_deleted() {
has_structure_changes = true;
- workspace_structure_change = Some((
- path,
- self.crate_graph_file_dependencies.contains(vfs_path),
- false,
- ));
+ workspace_structure_change =
+ Some((path, self.crate_graph_file_dependencies.contains(vfs_path)));
} else if path.extension() == Some("rs".as_ref()) {
modified_rust_files.push(file.file_id);
}
@@ -365,16 +358,11 @@ impl GlobalState {
// FIXME: ideally we should only trigger a workspace fetch for non-library changes
// but something's going wrong with the source root business when we add a new local
// crate see https://github.com/rust-lang/rust-analyzer/issues/13029
- if let Some((path, force_crate_graph_reload, build_scripts_touched)) =
- workspace_structure_change
- {
+ if let Some((path, force_crate_graph_reload)) = workspace_structure_change {
self.fetch_workspaces_queue.request_op(
format!("workspace vfs file change: {path}"),
force_crate_graph_reload,
);
- if build_scripts_touched {
- self.fetch_build_data_queue.request_op(format!("build.rs changed: {path}"), ());
- }
}
}
diff --git a/crates/rust-analyzer/src/handlers/request.rs b/crates/rust-analyzer/src/handlers/request.rs
index eb9d4bf0f0..04a0439542 100644
--- a/crates/rust-analyzer/src/handlers/request.rs
+++ b/crates/rust-analyzer/src/handlers/request.rs
@@ -16,6 +16,7 @@ use ide::{
ReferenceCategory, Runnable, RunnableKind, SingleResolve, SourceChange, TextEdit,
};
use ide_db::SymbolKind;
+use itertools::Itertools;
use lsp_server::ErrorCode;
use lsp_types::{
CallHierarchyIncomingCall, CallHierarchyIncomingCallsParams, CallHierarchyItem,
@@ -1055,9 +1056,8 @@ pub(crate) fn handle_references(
let exclude_imports = snap.config.find_all_refs_exclude_imports();
let exclude_tests = snap.config.find_all_refs_exclude_tests();
- let refs = match snap.analysis.find_all_refs(position, None)? {
- None => return Ok(None),
- Some(refs) => refs,
+ let Some(refs) = snap.analysis.find_all_refs(position, None)? else {
+ return Ok(None);
};
let include_declaration = params.context.include_declaration;
@@ -1084,6 +1084,7 @@ pub(crate) fn handle_references(
})
.chain(decl)
})
+ .unique()
.filter_map(|frange| to_proto::location(&snap, frange).ok())
.collect();
@@ -1802,10 +1803,10 @@ fn show_ref_command_link(
.into_iter()
.flat_map(|res| res.references)
.flat_map(|(file_id, ranges)| {
- ranges.into_iter().filter_map(move |(range, _)| {
- to_proto::location(snap, FileRange { file_id, range }).ok()
- })
+ ranges.into_iter().map(move |(range, _)| FileRange { file_id, range })
})
+ .unique()
+ .filter_map(|range| to_proto::location(snap, range).ok())
.collect();
let title = to_proto::reference_title(locations.len());
let command = to_proto::command::show_references(title, &uri, position, locations);
diff --git a/crates/rust-analyzer/src/lsp/to_proto.rs b/crates/rust-analyzer/src/lsp/to_proto.rs
index 727007bba0..481ebfefd4 100644
--- a/crates/rust-analyzer/src/lsp/to_proto.rs
+++ b/crates/rust-analyzer/src/lsp/to_proto.rs
@@ -904,15 +904,16 @@ pub(crate) fn goto_definition_response(
if snap.config.location_link() {
let links = targets
.into_iter()
+ .unique_by(|nav| (nav.file_id, nav.full_range, nav.focus_range))
.map(|nav| location_link(snap, src, nav))
.collect::<Cancellable<Vec<_>>>()?;
Ok(links.into())
} else {
let locations = targets
.into_iter()
- .map(|nav| {
- location(snap, FileRange { file_id: nav.file_id, range: nav.focus_or_full_range() })
- })
+ .map(|nav| FileRange { file_id: nav.file_id, range: nav.focus_or_full_range() })
+ .unique()
+ .map(|range| location(snap, range))
.collect::<Cancellable<Vec<_>>>()?;
Ok(locations.into())
}
@@ -1001,10 +1002,8 @@ fn merge_text_and_snippet_edits(
let mut new_text = current_indel.insert;
// find which snippet bits need to be escaped
- let escape_places = new_text
- .rmatch_indices(['\\', '$', '{', '}'])
- .map(|(insert, _)| insert)
- .collect_vec();
+ let escape_places =
+ new_text.rmatch_indices(['\\', '$', '}']).map(|(insert, _)| insert).collect_vec();
let mut escape_places = escape_places.into_iter().peekable();
let mut escape_prior_bits = |new_text: &mut String, up_to: usize| {
for before in escape_places.peeking_take_while(|insert| *insert >= up_to) {
@@ -2175,7 +2174,7 @@ fn bar(_: usize) {}
character: 0,
},
},
- new_text: "\\$${1:ab\\{\\}\\$c\\\\d}ef",
+ new_text: "\\$${1:ab{\\}\\$c\\\\d}ef",
insert_text_format: Some(
Snippet,
),
@@ -2271,7 +2270,7 @@ struct ProcMacro {
character: 5,
},
},
- new_text: "$0disabled = false;\n ProcMacro \\{\n disabled,\n \\}",
+ new_text: "$0disabled = false;\n ProcMacro {\n disabled,\n \\}",
insert_text_format: Some(
Snippet,
),
@@ -2335,7 +2334,7 @@ struct P {
character: 5,
},
},
- new_text: "$0disabled = false;\n ProcMacro \\{\n disabled,\n \\}",
+ new_text: "$0disabled = false;\n ProcMacro {\n disabled,\n \\}",
insert_text_format: Some(
Snippet,
),
@@ -2400,7 +2399,7 @@ struct ProcMacro {
character: 5,
},
},
- new_text: "${0:disabled} = false;\n ProcMacro \\{\n disabled,\n \\}",
+ new_text: "${0:disabled} = false;\n ProcMacro {\n disabled,\n \\}",
insert_text_format: Some(
Snippet,
),
@@ -2465,7 +2464,7 @@ struct P {
character: 5,
},
},
- new_text: "${0:disabled} = false;\n ProcMacro \\{\n disabled,\n \\}",
+ new_text: "${0:disabled} = false;\n ProcMacro {\n disabled,\n \\}",
insert_text_format: Some(
Snippet,
),
diff --git a/crates/rust-analyzer/src/lsp/utils.rs b/crates/rust-analyzer/src/lsp/utils.rs
index 10335cb145..800c0eee53 100644
--- a/crates/rust-analyzer/src/lsp/utils.rs
+++ b/crates/rust-analyzer/src/lsp/utils.rs
@@ -134,6 +134,7 @@ impl GlobalState {
let token = lsp_types::ProgressToken::String(
cancel_token.unwrap_or_else(|| format!("rustAnalyzer/{title}")),
);
+ tracing::debug!(?token, ?state, "report_progress {message:?}");
let work_done_progress = match state {
Progress::Begin => {
self.send_request::<lsp_types::request::WorkDoneProgressCreate>(
diff --git a/crates/rust-analyzer/src/reload.rs b/crates/rust-analyzer/src/reload.rs
index 5895459d1f..f6bc032c01 100644
--- a/crates/rust-analyzer/src/reload.rs
+++ b/crates/rust-analyzer/src/reload.rs
@@ -411,10 +411,7 @@ impl GlobalState {
if *force_reload_crate_graph {
self.recreate_crate_graph(cause);
}
- if self.build_deps_changed && self.config.run_build_scripts() {
- self.build_deps_changed = false;
- self.fetch_build_data_queue.request_op("build_deps_changed".to_owned(), ());
- }
+
// Current build scripts do not match the version of the active
// workspace, so there's nothing for us to update.
return;
@@ -424,7 +421,7 @@ impl GlobalState {
// Here, we completely changed the workspace (Cargo.toml edit), so
// we don't care about build-script results, they are stale.
- // FIXME: can we abort the build scripts here?
+ // FIXME: can we abort the build scripts here if they are already running?
self.workspaces = Arc::new(workspaces);
if self.config.run_build_scripts() {
@@ -525,13 +522,14 @@ impl GlobalState {
}
fn recreate_crate_graph(&mut self, cause: String) {
- {
+ // crate graph construction relies on these paths, record them so when one of them gets
+ // deleted or created we trigger a reconstruction of the crate graph
+ let mut crate_graph_file_dependencies = FxHashSet::default();
+
+ let (crate_graph, proc_macro_paths, layouts, toolchains) = {
// Create crate graph from all the workspaces
let vfs = &mut self.vfs.write().0;
let loader = &mut self.loader;
- // crate graph construction relies on these paths, record them so when one of them gets
- // deleted or created we trigger a reconstruction of the crate graph
- let mut crate_graph_file_dependencies = FxHashSet::default();
let load = |path: &AbsPath| {
let _p = tracing::span!(tracing::Level::DEBUG, "switch_workspaces::load").entered();
@@ -548,25 +546,24 @@ impl GlobalState {
}
};
- let (crate_graph, proc_macro_paths, layouts, toolchains) =
- ws_to_crate_graph(&self.workspaces, self.config.extra_env(), load);
-
- let mut change = Change::new();
- if self.config.expand_proc_macros() {
- change.set_proc_macros(
- crate_graph
- .iter()
- .map(|id| (id, Err("Proc-macros have not been built yet".to_owned())))
- .collect(),
- );
- self.fetch_proc_macros_queue.request_op(cause, proc_macro_paths);
- }
- change.set_crate_graph(crate_graph);
- change.set_target_data_layouts(layouts);
- change.set_toolchains(toolchains);
- self.analysis_host.apply_change(change);
- self.crate_graph_file_dependencies = crate_graph_file_dependencies;
+ ws_to_crate_graph(&self.workspaces, self.config.extra_env(), load)
+ };
+ let mut change = Change::new();
+ if self.config.expand_proc_macros() {
+ change.set_proc_macros(
+ crate_graph
+ .iter()
+ .map(|id| (id, Err("Proc-macros have not been built yet".to_owned())))
+ .collect(),
+ );
+ self.fetch_proc_macros_queue.request_op(cause, proc_macro_paths);
}
+ change.set_crate_graph(crate_graph);
+ change.set_target_data_layouts(layouts);
+ change.set_toolchains(toolchains);
+ self.analysis_host.apply_change(change);
+ self.crate_graph_file_dependencies = crate_graph_file_dependencies;
+
self.process_changes();
self.reload_flycheck();
}
diff --git a/crates/rust-analyzer/tests/slow-tests/support.rs b/crates/rust-analyzer/tests/slow-tests/support.rs
index 392a717020..dfd25abc70 100644
--- a/crates/rust-analyzer/tests/slow-tests/support.rs
+++ b/crates/rust-analyzer/tests/slow-tests/support.rs
@@ -243,7 +243,7 @@ impl Server {
to_string_pretty(actual_part).unwrap(),
);
} else {
- tracing::debug!("sucessfully matched notification");
+ tracing::debug!("successfully matched notification");
return;
}
} else {
diff --git a/crates/salsa/Cargo.toml b/crates/salsa/Cargo.toml
index 4ccbc3de84..9eec21f6a1 100644
--- a/crates/salsa/Cargo.toml
+++ b/crates/salsa/Cargo.toml
@@ -21,6 +21,7 @@ rustc-hash = "1.0"
smallvec = "1.0.0"
oorandom = "11"
triomphe = "0.1.11"
+itertools.workspace = true
salsa-macros = { version = "0.0.0", path = "salsa-macros" }
diff --git a/crates/salsa/salsa-macros/src/database_storage.rs b/crates/salsa/salsa-macros/src/database_storage.rs
index 0ec75bb043..223da9b529 100644
--- a/crates/salsa/salsa-macros/src/database_storage.rs
+++ b/crates/salsa/salsa-macros/src/database_storage.rs
@@ -154,8 +154,8 @@ pub(crate) fn database(args: TokenStream, input: TokenStream) -> TokenStream {
self.#db_storage_field.salsa_runtime()
}
- fn ops_salsa_runtime_mut(&mut self) -> &mut salsa::Runtime {
- self.#db_storage_field.salsa_runtime_mut()
+ fn synthetic_write(&mut self, durability: salsa::Durability) {
+ self.#db_storage_field.salsa_runtime_mut().synthetic_write(durability)
}
fn fmt_index(
diff --git a/crates/salsa/salsa-macros/src/query_group.rs b/crates/salsa/salsa-macros/src/query_group.rs
index 5d1678ef12..a868d920b6 100644
--- a/crates/salsa/salsa-macros/src/query_group.rs
+++ b/crates/salsa/salsa-macros/src/query_group.rs
@@ -526,7 +526,7 @@ pub(crate) fn query_group(args: TokenStream, input: TokenStream) -> TokenStream
fmt_ops.extend(quote! {
#query_index => {
salsa::plumbing::QueryStorageOps::fmt_index(
- &*self.#fn_name, db, input, fmt,
+ &*self.#fn_name, db, input.key_index(), fmt,
)
}
});
@@ -537,7 +537,7 @@ pub(crate) fn query_group(args: TokenStream, input: TokenStream) -> TokenStream
maybe_changed_ops.extend(quote! {
#query_index => {
salsa::plumbing::QueryStorageOps::maybe_changed_after(
- &*self.#fn_name, db, input, revision
+ &*self.#fn_name, db, input.key_index(), revision
)
}
});
diff --git a/crates/salsa/src/derived.rs b/crates/salsa/src/derived.rs
index d631671005..153df999f5 100644
--- a/crates/salsa/src/derived.rs
+++ b/crates/salsa/src/derived.rs
@@ -102,13 +102,13 @@ where
let mut write = self.slot_map.write();
let entry = write.entry(key.clone());
- let key_index = u32::try_from(entry.index()).unwrap();
+ let key_index = entry.index() as u32;
let database_key_index = DatabaseKeyIndex {
group_index: self.group_index,
query_index: Q::QUERY_INDEX,
key_index,
};
- entry.or_insert_with(|| Arc::new(Slot::new(key.clone(), database_key_index))).clone()
+ entry.or_insert_with(|| Arc::new(Slot::new(database_key_index))).clone()
}
}
@@ -131,34 +131,36 @@ where
fn fmt_index(
&self,
_db: &<Q as QueryDb<'_>>::DynDb,
- index: DatabaseKeyIndex,
+ index: u32,
fmt: &mut std::fmt::Formatter<'_>,
) -> std::fmt::Result {
- assert_eq!(index.group_index, self.group_index);
- assert_eq!(index.query_index, Q::QUERY_INDEX);
let slot_map = self.slot_map.read();
- let key = slot_map.get_index(index.key_index as usize).unwrap().0;
+ let key = slot_map.get_index(index as usize).unwrap().0;
write!(fmt, "{}({:?})", Q::QUERY_NAME, key)
}
fn maybe_changed_after(
&self,
db: &<Q as QueryDb<'_>>::DynDb,
- input: DatabaseKeyIndex,
+ index: u32,
revision: Revision,
) -> bool {
- assert_eq!(input.group_index, self.group_index);
- assert_eq!(input.query_index, Q::QUERY_INDEX);
debug_assert!(revision < db.salsa_runtime().current_revision());
- let slot = self.slot_map.read().get_index(input.key_index as usize).unwrap().1.clone();
- slot.maybe_changed_after(db, revision)
+ let read = self.slot_map.read();
+ let Some((key, slot)) = read.get_index(index as usize) else {
+ return false;
+ };
+ let (key, slot) = (key.clone(), slot.clone());
+ // note: this drop is load-bearing. removing it would causes deadlocks.
+ drop(read);
+ slot.maybe_changed_after(db, revision, &key)
}
fn fetch(&self, db: &<Q as QueryDb<'_>>::DynDb, key: &Q::Key) -> Q::Value {
db.unwind_if_cancelled();
let slot = self.slot(key);
- let StampedValue { value, durability, changed_at } = slot.read(db);
+ let StampedValue { value, durability, changed_at } = slot.read(db, key);
if let Some(evicted) = self.lru_list.record_use(&slot) {
evicted.evict();
@@ -182,7 +184,7 @@ where
C: std::iter::FromIterator<TableEntry<Q::Key, Q::Value>>,
{
let slot_map = self.slot_map.read();
- slot_map.values().filter_map(|slot| slot.as_table_entry()).collect()
+ slot_map.iter().filter_map(|(key, slot)| slot.as_table_entry(key)).collect()
}
}
diff --git a/crates/salsa/src/derived/slot.rs b/crates/salsa/src/derived/slot.rs
index 4fad791a26..75204c8ff6 100644
--- a/crates/salsa/src/derived/slot.rs
+++ b/crates/salsa/src/derived/slot.rs
@@ -26,8 +26,8 @@ where
Q: QueryFunction,
MP: MemoizationPolicy<Q>,
{
- key: Q::Key,
- database_key_index: DatabaseKeyIndex,
+ key_index: u32,
+ group_index: u16,
state: RwLock<QueryState<Q>>,
policy: PhantomData<MP>,
lru_index: LruIndex,
@@ -110,10 +110,10 @@ where
Q: QueryFunction,
MP: MemoizationPolicy<Q>,
{
- pub(super) fn new(key: Q::Key, database_key_index: DatabaseKeyIndex) -> Self {
+ pub(super) fn new(database_key_index: DatabaseKeyIndex) -> Self {
Self {
- key,
- database_key_index,
+ key_index: database_key_index.key_index,
+ group_index: database_key_index.group_index,
state: RwLock::new(QueryState::NotComputed),
lru_index: LruIndex::default(),
policy: PhantomData,
@@ -121,10 +121,18 @@ where
}
pub(super) fn database_key_index(&self) -> DatabaseKeyIndex {
- self.database_key_index
+ DatabaseKeyIndex {
+ group_index: self.group_index,
+ query_index: Q::QUERY_INDEX,
+ key_index: self.key_index,
+ }
}
- pub(super) fn read(&self, db: &<Q as QueryDb<'_>>::DynDb) -> StampedValue<Q::Value> {
+ pub(super) fn read(
+ &self,
+ db: &<Q as QueryDb<'_>>::DynDb,
+ key: &Q::Key,
+ ) -> StampedValue<Q::Value> {
let runtime = db.salsa_runtime();
// NB: We don't need to worry about people modifying the
@@ -147,7 +155,7 @@ where
}
}
- self.read_upgrade(db, revision_now)
+ self.read_upgrade(db, key, revision_now)
}
/// Second phase of a read operation: acquires an upgradable-read
@@ -157,6 +165,7 @@ where
fn read_upgrade(
&self,
db: &<Q as QueryDb<'_>>::DynDb,
+ key: &Q::Key,
revision_now: Revision,
) -> StampedValue<Q::Value> {
let runtime = db.salsa_runtime();
@@ -186,8 +195,8 @@ where
}
};
- let panic_guard = PanicGuard::new(self.database_key_index, self, runtime);
- let active_query = runtime.push_query(self.database_key_index);
+ let panic_guard = PanicGuard::new(self, runtime);
+ let active_query = runtime.push_query(self.database_key_index());
// If we have an old-value, it *may* now be stale, since there
// has been a new revision since the last time we checked. So,
@@ -200,7 +209,7 @@ where
db.salsa_event(Event {
runtime_id: runtime.id(),
kind: EventKind::DidValidateMemoizedValue {
- database_key: self.database_key_index,
+ database_key: self.database_key_index(),
},
});
@@ -210,7 +219,7 @@ where
}
}
- self.execute(db, runtime, revision_now, active_query, panic_guard, old_memo)
+ self.execute(db, runtime, revision_now, active_query, panic_guard, old_memo, key)
}
fn execute(
@@ -221,22 +230,23 @@ where
active_query: ActiveQueryGuard<'_>,
panic_guard: PanicGuard<'_, Q, MP>,
old_memo: Option<Memo<Q::Value>>,
+ key: &Q::Key,
) -> StampedValue<Q::Value> {
- tracing::info!("{:?}: executing query", self.database_key_index.debug(db));
+ tracing::info!("{:?}: executing query", self.database_key_index().debug(db));
db.salsa_event(Event {
runtime_id: db.salsa_runtime().id(),
- kind: EventKind::WillExecute { database_key: self.database_key_index },
+ kind: EventKind::WillExecute { database_key: self.database_key_index() },
});
// Query was not previously executed, or value is potentially
// stale, or value is absent. Let's execute!
- let value = match Cycle::catch(|| Q::execute(db, self.key.clone())) {
+ let value = match Cycle::catch(|| Q::execute(db, key.clone())) {
Ok(v) => v,
Err(cycle) => {
tracing::debug!(
"{:?}: caught cycle {:?}, have strategy {:?}",
- self.database_key_index.debug(db),
+ self.database_key_index().debug(db),
cycle,
Q::CYCLE_STRATEGY,
);
@@ -248,12 +258,12 @@ where
crate::plumbing::CycleRecoveryStrategy::Fallback => {
if let Some(c) = active_query.take_cycle() {
assert!(c.is(&cycle));
- Q::cycle_fallback(db, &cycle, &self.key)
+ Q::cycle_fallback(db, &cycle, key)
} else {
// we are not a participant in this cycle
debug_assert!(!cycle
.participant_keys()
- .any(|k| k == self.database_key_index));
+ .any(|k| k == self.database_key_index()));
cycle.throw()
}
}
@@ -303,7 +313,7 @@ where
};
let memo_value =
- if self.should_memoize_value(&self.key) { Some(new_value.value.clone()) } else { None };
+ if self.should_memoize_value(key) { Some(new_value.value.clone()) } else { None };
debug!("read_upgrade({:?}): result.revisions = {:#?}", self, revisions,);
@@ -395,13 +405,11 @@ where
}
}
- pub(super) fn as_table_entry(&self) -> Option<TableEntry<Q::Key, Q::Value>> {
+ pub(super) fn as_table_entry(&self, key: &Q::Key) -> Option<TableEntry<Q::Key, Q::Value>> {
match &*self.state.read() {
QueryState::NotComputed => None,
- QueryState::InProgress { .. } => Some(TableEntry::new(self.key.clone(), None)),
- QueryState::Memoized(memo) => {
- Some(TableEntry::new(self.key.clone(), memo.value.clone()))
- }
+ QueryState::InProgress { .. } => Some(TableEntry::new(key.clone(), None)),
+ QueryState::Memoized(memo) => Some(TableEntry::new(key.clone(), memo.value.clone())),
}
}
@@ -436,6 +444,7 @@ where
&self,
db: &<Q as QueryDb<'_>>::DynDb,
revision: Revision,
+ key: &Q::Key,
) -> bool {
let runtime = db.salsa_runtime();
let revision_now = runtime.current_revision();
@@ -458,7 +467,7 @@ where
MaybeChangedSinceProbeState::ChangedAt(changed_at) => return changed_at > revision,
MaybeChangedSinceProbeState::Stale(state) => {
drop(state);
- return self.maybe_changed_after_upgrade(db, revision);
+ return self.maybe_changed_after_upgrade(db, revision, key);
}
}
}
@@ -495,6 +504,7 @@ where
&self,
db: &<Q as QueryDb<'_>>::DynDb,
revision: Revision,
+ key: &Q::Key,
) -> bool {
let runtime = db.salsa_runtime();
let revision_now = runtime.current_revision();
@@ -513,7 +523,9 @@ where
// If another thread was active, then the cache line is going to be
// either verified or cleared out. Just recurse to figure out which.
// Note that we don't need an upgradable read.
- MaybeChangedSinceProbeState::Retry => return self.maybe_changed_after(db, revision),
+ MaybeChangedSinceProbeState::Retry => {
+ return self.maybe_changed_after(db, revision, key)
+ }
MaybeChangedSinceProbeState::Stale(state) => {
type RwLockUpgradableReadGuard<'a, T> =
@@ -527,8 +539,8 @@ where
}
};
- let panic_guard = PanicGuard::new(self.database_key_index, self, runtime);
- let active_query = runtime.push_query(self.database_key_index);
+ let panic_guard = PanicGuard::new(self, runtime);
+ let active_query = runtime.push_query(self.database_key_index());
if old_memo.verify_revisions(db.ops_database(), revision_now, &active_query) {
let maybe_changed = old_memo.revisions.changed_at > revision;
@@ -538,8 +550,15 @@ where
// We found that this memoized value may have changed
// but we have an old value. We can re-run the code and
// actually *check* if it has changed.
- let StampedValue { changed_at, .. } =
- self.execute(db, runtime, revision_now, active_query, panic_guard, Some(old_memo));
+ let StampedValue { changed_at, .. } = self.execute(
+ db,
+ runtime,
+ revision_now,
+ active_query,
+ panic_guard,
+ Some(old_memo),
+ key,
+ );
changed_at > revision
} else {
// We found that inputs to this memoized value may have chanced
@@ -560,7 +579,7 @@ where
) {
runtime.block_on_or_unwind(
db.ops_database(),
- self.database_key_index,
+ self.database_key_index(),
other_id,
mutex_guard,
)
@@ -585,7 +604,6 @@ where
Q: QueryFunction,
MP: MemoizationPolicy<Q>,
{
- database_key_index: DatabaseKeyIndex,
slot: &'me Slot<Q, MP>,
runtime: &'me Runtime,
}
@@ -595,12 +613,8 @@ where
Q: QueryFunction,
MP: MemoizationPolicy<Q>,
{
- fn new(
- database_key_index: DatabaseKeyIndex,
- slot: &'me Slot<Q, MP>,
- runtime: &'me Runtime,
- ) -> Self {
- Self { database_key_index, slot, runtime }
+ fn new(slot: &'me Slot<Q, MP>, runtime: &'me Runtime) -> Self {
+ Self { slot, runtime }
}
/// Indicates that we have concluded normally (without panicking).
@@ -616,17 +630,18 @@ where
/// inserted; if others were blocked, waiting for us to finish,
/// then notify them.
fn overwrite_placeholder(&mut self, wait_result: WaitResult, opt_memo: Option<Memo<Q::Value>>) {
- let mut write = self.slot.state.write();
-
- let old_value = match opt_memo {
- // Replace the `InProgress` marker that we installed with the new
- // memo, thus releasing our unique access to this key.
- Some(memo) => std::mem::replace(&mut *write, QueryState::Memoized(memo)),
-
- // We had installed an `InProgress` marker, but we panicked before
- // it could be removed. At this point, we therefore "own" unique
- // access to our slot, so we can just remove the key.
- None => std::mem::replace(&mut *write, QueryState::NotComputed),
+ let old_value = {
+ let mut write = self.slot.state.write();
+ match opt_memo {
+ // Replace the `InProgress` marker that we installed with the new
+ // memo, thus releasing our unique access to this key.
+ Some(memo) => std::mem::replace(&mut *write, QueryState::Memoized(memo)),
+
+ // We had installed an `InProgress` marker, but we panicked before
+ // it could be removed. At this point, we therefore "own" unique
+ // access to our slot, so we can just remove the key.
+ None => std::mem::replace(&mut *write, QueryState::NotComputed),
+ }
};
match old_value {
@@ -638,7 +653,8 @@ where
// acquire a mutex; the mutex will guarantee that all writes
// we are interested in are visible.
if anyone_waiting.load(Ordering::Relaxed) {
- self.runtime.unblock_queries_blocked_on(self.database_key_index, wait_result);
+ self.runtime
+ .unblock_queries_blocked_on(self.slot.database_key_index(), wait_result);
}
}
_ => panic!(
@@ -692,10 +708,10 @@ where
return None;
}
if self.verify_revisions(db, revision_now, active_query) {
- Some(StampedValue {
+ self.value.clone().map(|value| StampedValue {
durability: self.revisions.durability,
changed_at: self.revisions.changed_at,
- value: self.value.as_ref().unwrap().clone(),
+ value,
})
} else {
None
@@ -748,7 +764,7 @@ where
// input changed *again*.
QueryInputs::Tracked { inputs } => {
let changed_input =
- inputs.iter().find(|&&input| db.maybe_changed_after(input, verified_at));
+ inputs.slice.iter().find(|&&input| db.maybe_changed_after(input, verified_at));
if let Some(input) = changed_input {
debug!("validate_memoized_value: `{:?}` may have changed", input);
@@ -788,7 +804,7 @@ where
MP: MemoizationPolicy<Q>,
{
fn fmt(&self, fmt: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
- write!(fmt, "{:?}({:?})", Q::default(), self.key)
+ write!(fmt, "{:?}", Q::default())
}
}
diff --git a/crates/salsa/src/durability.rs b/crates/salsa/src/durability.rs
index 0c82f6345a..44abae3170 100644
--- a/crates/salsa/src/durability.rs
+++ b/crates/salsa/src/durability.rs
@@ -42,9 +42,9 @@ impl Durability {
pub(crate) const MAX: Durability = Self::HIGH;
/// Number of durability levels.
- pub(crate) const LEN: usize = 3;
+ pub(crate) const LEN: usize = Self::MAX.index() + 1;
- pub(crate) fn index(self) -> usize {
+ pub(crate) const fn index(self) -> usize {
self.0 as usize
}
}
diff --git a/crates/salsa/src/input.rs b/crates/salsa/src/input.rs
index c2539570e0..922ec5a775 100644
--- a/crates/salsa/src/input.rs
+++ b/crates/salsa/src/input.rs
@@ -29,7 +29,7 @@ where
}
struct Slot<V> {
- database_key_index: DatabaseKeyIndex,
+ key_index: u32,
stamped_value: RwLock<StampedValue<V>>,
}
@@ -54,27 +54,25 @@ where
fn fmt_index(
&self,
_db: &<Q as QueryDb<'_>>::DynDb,
- index: DatabaseKeyIndex,
+ index: u32,
fmt: &mut std::fmt::Formatter<'_>,
) -> std::fmt::Result {
- assert_eq!(index.group_index, self.group_index);
- assert_eq!(index.query_index, Q::QUERY_INDEX);
let slot_map = self.slots.read();
- let key = slot_map.get_index(index.key_index as usize).unwrap().0;
+ let key = slot_map.get_index(index as usize).unwrap().0;
write!(fmt, "{}({:?})", Q::QUERY_NAME, key)
}
fn maybe_changed_after(
&self,
db: &<Q as QueryDb<'_>>::DynDb,
- input: DatabaseKeyIndex,
+ index: u32,
revision: Revision,
) -> bool {
- assert_eq!(input.group_index, self.group_index);
- assert_eq!(input.query_index, Q::QUERY_INDEX);
debug_assert!(revision < db.salsa_runtime().current_revision());
let slots = &self.slots.read();
- let slot = slots.get_index(input.key_index as usize).unwrap().1;
+ let Some((_, slot)) = slots.get_index(index as usize) else {
+ return true;
+ };
debug!("maybe_changed_after(slot={:?}, revision={:?})", Q::default(), revision,);
@@ -96,7 +94,11 @@ where
let StampedValue { value, durability, changed_at } = slot.stamped_value.read().clone();
db.salsa_runtime().report_query_read_and_unwind_if_cycle_resulted(
- slot.database_key_index,
+ DatabaseKeyIndex {
+ group_index: self.group_index,
+ query_index: Q::QUERY_INDEX,
+ key_index: slot.key_index,
+ },
durability,
changed_at,
);
@@ -174,16 +176,8 @@ where
}
Entry::Vacant(entry) => {
- let key_index = u32::try_from(entry.index()).unwrap();
- let database_key_index = DatabaseKeyIndex {
- group_index: self.group_index,
- query_index: Q::QUERY_INDEX,
- key_index,
- };
- entry.insert(Slot {
- database_key_index,
- stamped_value: RwLock::new(stamped_value),
- });
+ let key_index = entry.index() as u32;
+ entry.insert(Slot { key_index, stamped_value: RwLock::new(stamped_value) });
None
}
}
@@ -196,7 +190,6 @@ pub struct UnitInputStorage<Q>
where
Q: Query<Key = ()>,
{
- group_index: u16,
slot: UnitSlot<Q::Value>,
}
@@ -222,36 +215,32 @@ where
fn new(group_index: u16) -> Self {
let database_key_index =
DatabaseKeyIndex { group_index, query_index: Q::QUERY_INDEX, key_index: 0 };
- UnitInputStorage {
- group_index,
- slot: UnitSlot { database_key_index, stamped_value: RwLock::new(None) },
- }
+ UnitInputStorage { slot: UnitSlot { database_key_index, stamped_value: RwLock::new(None) } }
}
fn fmt_index(
&self,
_db: &<Q as QueryDb<'_>>::DynDb,
- index: DatabaseKeyIndex,
+ _index: u32,
fmt: &mut std::fmt::Formatter<'_>,
) -> std::fmt::Result {
- assert_eq!(index.group_index, self.group_index);
- assert_eq!(index.query_index, Q::QUERY_INDEX);
write!(fmt, "{}", Q::QUERY_NAME)
}
fn maybe_changed_after(
&self,
db: &<Q as QueryDb<'_>>::DynDb,
- input: DatabaseKeyIndex,
+ _index: u32,
revision: Revision,
) -> bool {
- assert_eq!(input.group_index, self.group_index);
- assert_eq!(input.query_index, Q::QUERY_INDEX);
debug_assert!(revision < db.salsa_runtime().current_revision());
debug!("maybe_changed_after(slot={:?}, revision={:?})", Q::default(), revision,);
- let changed_at = self.slot.stamped_value.read().as_ref().unwrap().changed_at;
+ let Some(value) = &*self.slot.stamped_value.read() else {
+ return true;
+ };
+ let changed_at = value.changed_at;
debug!("maybe_changed_after: changed_at = {:?}", changed_at);
diff --git a/crates/salsa/src/interned.rs b/crates/salsa/src/interned.rs
index 822219f518..c065e7e2bd 100644
--- a/crates/salsa/src/interned.rs
+++ b/crates/salsa/src/interned.rs
@@ -265,12 +265,10 @@ where
fn fmt_index(
&self,
_db: &<Q as QueryDb<'_>>::DynDb,
- index: DatabaseKeyIndex,
+ index: u32,
fmt: &mut std::fmt::Formatter<'_>,
) -> std::fmt::Result {
- assert_eq!(index.group_index, self.group_index);
- assert_eq!(index.query_index, Q::QUERY_INDEX);
- let intern_id = InternId::from(index.key_index);
+ let intern_id = InternId::from(index);
let slot = self.lookup_value(intern_id);
write!(fmt, "{}({:?})", Q::QUERY_NAME, slot.value)
}
@@ -278,13 +276,11 @@ where
fn maybe_changed_after(
&self,
db: &<Q as QueryDb<'_>>::DynDb,
- input: DatabaseKeyIndex,
+ input: u32,
revision: Revision,
) -> bool {
- assert_eq!(input.group_index, self.group_index);
- assert_eq!(input.query_index, Q::QUERY_INDEX);
debug_assert!(revision < db.salsa_runtime().current_revision());
- let intern_id = InternId::from(input.key_index);
+ let intern_id = InternId::from(input);
let slot = self.lookup_value(intern_id);
slot.maybe_changed_after(revision)
}
@@ -388,7 +384,7 @@ where
fn fmt_index(
&self,
db: &<Q as QueryDb<'_>>::DynDb,
- index: DatabaseKeyIndex,
+ index: u32,
fmt: &mut std::fmt::Formatter<'_>,
) -> std::fmt::Result {
let group_storage =
@@ -400,7 +396,7 @@ where
fn maybe_changed_after(
&self,
db: &<Q as QueryDb<'_>>::DynDb,
- input: DatabaseKeyIndex,
+ input: u32,
revision: Revision,
) -> bool {
let group_storage =
diff --git a/crates/salsa/src/lib.rs b/crates/salsa/src/lib.rs
index 668dcfd925..fe80759887 100644
--- a/crates/salsa/src/lib.rs
+++ b/crates/salsa/src/lib.rs
@@ -54,7 +54,7 @@ pub trait Database: plumbing::DatabaseOps {
/// runtime. It permits the database to be customized and to
/// inject logging or other custom behavior.
fn salsa_event(&self, event_fn: Event) {
- #![allow(unused_variables)]
+ _ = event_fn;
}
/// Starts unwinding the stack if the current revision is cancelled.
@@ -96,11 +96,16 @@ pub trait Database: plumbing::DatabaseOps {
self.ops_salsa_runtime()
}
- /// Gives access to the underlying salsa runtime.
+ /// A "synthetic write" causes the system to act *as though* some
+ /// input of durability `durability` has changed. This is mostly
+ /// useful for profiling scenarios.
///
- /// This method should not be overridden by `Database` implementors.
- fn salsa_runtime_mut(&mut self) -> &mut Runtime {
- self.ops_salsa_runtime_mut()
+ /// **WARNING:** Just like an ordinary write, this method triggers
+ /// cancellation. If you invoke it while a snapshot exists, it
+ /// will block until that snapshot is dropped -- if that snapshot
+ /// is owned by the current thread, this could trigger deadlock.
+ fn synthetic_write(&mut self, durability: Durability) {
+ plumbing::DatabaseOps::synthetic_write(self, durability)
}
}
@@ -456,12 +461,12 @@ pub trait Query: Debug + Default + Sized + for<'d> QueryDb<'d> {
/// Name of the query method (e.g., `foo`)
const QUERY_NAME: &'static str;
- /// Extact storage for this query from the storage for its group.
+ /// Extract storage for this query from the storage for its group.
fn query_storage<'a>(
group_storage: &'a <Self as QueryDb<'_>>::GroupStorage,
) -> &'a std::sync::Arc<Self::Storage>;
- /// Extact storage for this query from the storage for its group.
+ /// Extract storage for this query from the storage for its group.
fn query_storage_mut<'a>(
group_storage: &'a <Self as QueryDb<'_>>::GroupStorage,
) -> &'a std::sync::Arc<Self::Storage>;
diff --git a/crates/salsa/src/lru.rs b/crates/salsa/src/lru.rs
index c6b9778f20..1ff85a3ea4 100644
--- a/crates/salsa/src/lru.rs
+++ b/crates/salsa/src/lru.rs
@@ -40,7 +40,7 @@ pub(crate) trait LruNode: Sized + Debug {
#[derive(Debug)]
pub(crate) struct LruIndex {
- /// Index in the approprate LRU list, or std::usize::MAX if not a
+ /// Index in the appropriate LRU list, or std::usize::MAX if not a
/// member.
index: AtomicUsize,
}
diff --git a/crates/salsa/src/plumbing.rs b/crates/salsa/src/plumbing.rs
index 71332e39ca..1a8ff33b2e 100644
--- a/crates/salsa/src/plumbing.rs
+++ b/crates/salsa/src/plumbing.rs
@@ -38,8 +38,15 @@ pub trait DatabaseOps {
/// Gives access to the underlying salsa runtime.
fn ops_salsa_runtime(&self) -> &Runtime;
- /// Gives access to the underlying salsa runtime.
- fn ops_salsa_runtime_mut(&mut self) -> &mut Runtime;
+ /// A "synthetic write" causes the system to act *as though* some
+ /// input of durability `durability` has changed. This is mostly
+ /// useful for profiling scenarios.
+ ///
+ /// **WARNING:** Just like an ordinary write, this method triggers
+ /// cancellation. If you invoke it while a snapshot exists, it
+ /// will block until that snapshot is dropped -- if that snapshot
+ /// is owned by the current thread, this could trigger deadlock.
+ fn synthetic_write(&mut self, durability: Durability);
/// Formats a database key index in a human readable fashion.
fn fmt_index(
@@ -166,7 +173,7 @@ where
fn fmt_index(
&self,
db: &<Q as QueryDb<'_>>::DynDb,
- index: DatabaseKeyIndex,
+ index: u32,
fmt: &mut std::fmt::Formatter<'_>,
) -> std::fmt::Result;
@@ -179,7 +186,7 @@ where
fn maybe_changed_after(
&self,
db: &<Q as QueryDb<'_>>::DynDb,
- input: DatabaseKeyIndex,
+ index: u32,
revision: Revision,
) -> bool;
// ANCHOR_END:maybe_changed_after
diff --git a/crates/salsa/src/revision.rs b/crates/salsa/src/revision.rs
index d97aaf9deb..559b033860 100644
--- a/crates/salsa/src/revision.rs
+++ b/crates/salsa/src/revision.rs
@@ -46,7 +46,7 @@ pub(crate) struct AtomicRevision {
}
impl AtomicRevision {
- pub(crate) fn start() -> Self {
+ pub(crate) const fn start() -> Self {
Self { data: AtomicU32::new(START) }
}
diff --git a/crates/salsa/src/runtime.rs b/crates/salsa/src/runtime.rs
index 40b8856991..a7d5a24578 100644
--- a/crates/salsa/src/runtime.rs
+++ b/crates/salsa/src/runtime.rs
@@ -4,13 +4,14 @@ use crate::hash::FxIndexSet;
use crate::plumbing::CycleRecoveryStrategy;
use crate::revision::{AtomicRevision, Revision};
use crate::{Cancelled, Cycle, Database, DatabaseKeyIndex, Event, EventKind};
+use itertools::Itertools;
use parking_lot::lock_api::{RawRwLock, RawRwLockRecursive};
use parking_lot::{Mutex, RwLock};
use std::hash::Hash;
use std::panic::panic_any;
-use std::sync::atomic::{AtomicUsize, Ordering};
+use std::sync::atomic::{AtomicU32, Ordering};
use tracing::debug;
-use triomphe::Arc;
+use triomphe::{Arc, ThinArc};
mod dependency_graph;
use dependency_graph::DependencyGraph;
@@ -297,8 +298,7 @@ impl Runtime {
// (at least for this execution, not necessarily across executions),
// no matter where it started on the stack. Find the minimum
// key and rotate it to the front.
- let min = v.iter().min().unwrap();
- let index = v.iter().position(|p| p == min).unwrap();
+ let index = v.iter().position_min().unwrap_or_default();
v.rotate_left(index);
// No need to store extra memory.
@@ -440,7 +440,7 @@ impl Runtime {
/// State that will be common to all threads (when we support multiple threads)
struct SharedState {
/// Stores the next id to use for a snapshotted runtime (starts at 1).
- next_id: AtomicUsize,
+ next_id: AtomicU32,
/// Whenever derived queries are executing, they acquire this lock
/// in read mode. Mutating inputs (and thus creating a new
@@ -457,50 +457,46 @@ struct SharedState {
/// revision is cancelled).
pending_revision: AtomicRevision,
- /// Stores the "last change" revision for values of each duration.
+ /// Stores the "last change" revision for values of each Durability.
/// This vector is always of length at least 1 (for Durability 0)
- /// but its total length depends on the number of durations. The
+ /// but its total length depends on the number of Durabilities. The
/// element at index 0 is special as it represents the "current
/// revision". In general, we have the invariant that revisions
/// in here are *declining* -- that is, `revisions[i] >=
/// revisions[i + 1]`, for all `i`. This is because when you
/// modify a value with durability D, that implies that values
/// with durability less than D may have changed too.
- revisions: Vec<AtomicRevision>,
+ revisions: [AtomicRevision; Durability::LEN],
/// The dependency graph tracks which runtimes are blocked on one
/// another, waiting for queries to terminate.
dependency_graph: Mutex<DependencyGraph>,
}
-impl SharedState {
- fn with_durabilities(durabilities: usize) -> Self {
- SharedState {
- next_id: AtomicUsize::new(1),
- query_lock: Default::default(),
- revisions: (0..durabilities).map(|_| AtomicRevision::start()).collect(),
- pending_revision: AtomicRevision::start(),
- dependency_graph: Default::default(),
- }
- }
-}
-
impl std::panic::RefUnwindSafe for SharedState {}
impl Default for SharedState {
fn default() -> Self {
- Self::with_durabilities(Durability::LEN)
+ #[allow(clippy::declare_interior_mutable_const)]
+ const START: AtomicRevision = AtomicRevision::start();
+ SharedState {
+ next_id: AtomicU32::new(1),
+ query_lock: Default::default(),
+ revisions: [START; Durability::LEN],
+ pending_revision: START,
+ dependency_graph: Default::default(),
+ }
}
}
impl std::fmt::Debug for SharedState {
fn fmt(&self, fmt: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
- let query_lock = if self.query_lock.try_write().is_some() {
- "<unlocked>"
- } else if self.query_lock.try_read().is_some() {
+ let query_lock = if self.query_lock.is_locked_exclusive() {
+ "<wlocked>"
+ } else if self.query_lock.is_locked() {
"<rlocked>"
} else {
- "<wlocked>"
+ "<unlocked>"
};
fmt.debug_struct("SharedState")
.field("query_lock", &query_lock)
@@ -570,7 +566,9 @@ impl ActiveQuery {
if dependencies.is_empty() {
QueryInputs::NoInputs
} else {
- QueryInputs::Tracked { inputs: dependencies.iter().copied().collect() }
+ QueryInputs::Tracked {
+ inputs: ThinArc::from_header_and_iter((), dependencies.iter().copied()),
+ }
}
}
};
@@ -616,7 +614,7 @@ impl ActiveQuery {
/// complete, its `RuntimeId` may potentially be re-used.
#[derive(Copy, Clone, Debug, PartialEq, Eq, Hash, PartialOrd, Ord)]
pub struct RuntimeId {
- counter: usize,
+ counter: u32,
}
#[derive(Clone, Debug)]
diff --git a/crates/salsa/src/runtime/dependency_graph.rs b/crates/salsa/src/runtime/dependency_graph.rs
index e41eb280de..dd223eeeba 100644
--- a/crates/salsa/src/runtime/dependency_graph.rs
+++ b/crates/salsa/src/runtime/dependency_graph.rs
@@ -12,7 +12,7 @@ type QueryStack = Vec<ActiveQuery>;
#[derive(Debug, Default)]
pub(super) struct DependencyGraph {
- /// A `(K -> V)` pair in this map indicates that the the runtime
+ /// A `(K -> V)` pair in this map indicates that the runtime
/// `K` is blocked on some query executing in the runtime `V`.
/// This encodes a graph that must be acyclic (or else deadlock
/// will result).
diff --git a/crates/salsa/src/runtime/local_state.rs b/crates/salsa/src/runtime/local_state.rs
index 91b95dffe7..7ac21dec1a 100644
--- a/crates/salsa/src/runtime/local_state.rs
+++ b/crates/salsa/src/runtime/local_state.rs
@@ -1,5 +1,6 @@
//!
use tracing::debug;
+use triomphe::ThinArc;
use crate::durability::Durability;
use crate::runtime::ActiveQuery;
@@ -7,7 +8,6 @@ use crate::runtime::Revision;
use crate::Cycle;
use crate::DatabaseKeyIndex;
use std::cell::RefCell;
-use triomphe::Arc;
/// State that is specific to a single execution thread.
///
@@ -43,7 +43,7 @@ pub(crate) struct QueryRevisions {
#[derive(Debug, Clone)]
pub(crate) enum QueryInputs {
/// Non-empty set of inputs, fully known
- Tracked { inputs: Arc<[DatabaseKeyIndex]> },
+ Tracked { inputs: ThinArc<(), DatabaseKeyIndex> },
/// Empty set of inputs, fully known.
NoInputs,
@@ -145,8 +145,7 @@ impl LocalState {
/// the current thread is blocking. The stack must be restored
/// with [`Self::restore_query_stack`] when the thread unblocks.
pub(super) fn take_query_stack(&self) -> Vec<ActiveQuery> {
- assert!(self.query_stack.borrow().is_some(), "query stack already taken");
- self.query_stack.take().unwrap()
+ self.query_stack.take().expect("query stack already taken")
}
/// Restores a query stack taken with [`Self::take_query_stack`] once
diff --git a/crates/salsa/tests/incremental/memoized_volatile.rs b/crates/salsa/tests/incremental/memoized_volatile.rs
index 6dc5030063..3dcc32eece 100644
--- a/crates/salsa/tests/incremental/memoized_volatile.rs
+++ b/crates/salsa/tests/incremental/memoized_volatile.rs
@@ -58,7 +58,7 @@ fn revalidate() {
// Second generation: volatile will change (to 1) but memoized1
// will not (still 0, as 1/2 = 0)
- query.salsa_runtime_mut().synthetic_write(Durability::LOW);
+ query.synthetic_write(Durability::LOW);
query.memoized2();
query.assert_log(&["Volatile invoked", "Memoized1 invoked"]);
query.memoized2();
@@ -67,7 +67,7 @@ fn revalidate() {
// Third generation: volatile will change (to 2) and memoized1
// will too (to 1). Therefore, after validating that Memoized1
// changed, we now invoke Memoized2.
- query.salsa_runtime_mut().synthetic_write(Durability::LOW);
+ query.synthetic_write(Durability::LOW);
query.memoized2();
query.assert_log(&["Volatile invoked", "Memoized1 invoked", "Memoized2 invoked"]);
diff --git a/crates/salsa/tests/on_demand_inputs.rs b/crates/salsa/tests/on_demand_inputs.rs
index 5d0e486644..677d633ee7 100644
--- a/crates/salsa/tests/on_demand_inputs.rs
+++ b/crates/salsa/tests/on_demand_inputs.rs
@@ -111,7 +111,7 @@ fn on_demand_input_durability() {
}
"#]].assert_debug_eq(&events);
- db.salsa_runtime_mut().synthetic_write(Durability::LOW);
+ db.synthetic_write(Durability::LOW);
events.replace(vec![]);
assert_eq!(db.c(1), 10);
assert_eq!(db.c(2), 20);
@@ -128,7 +128,7 @@ fn on_demand_input_durability() {
}
"#]].assert_debug_eq(&events);
- db.salsa_runtime_mut().synthetic_write(Durability::HIGH);
+ db.synthetic_write(Durability::HIGH);
events.replace(vec![]);
assert_eq!(db.c(1), 10);
assert_eq!(db.c(2), 20);
diff --git a/crates/salsa/tests/storage_varieties/tests.rs b/crates/salsa/tests/storage_varieties/tests.rs
index f75c7c142f..8e2f9b03cb 100644
--- a/crates/salsa/tests/storage_varieties/tests.rs
+++ b/crates/salsa/tests/storage_varieties/tests.rs
@@ -20,7 +20,7 @@ fn volatile_twice() {
let v2 = db.volatile(); // volatiles are cached, so 2nd read returns the same
assert_eq!(v1, v2);
- db.salsa_runtime_mut().synthetic_write(Durability::LOW); // clears volatile caches
+ db.synthetic_write(Durability::LOW); // clears volatile caches
let v3 = db.volatile(); // will re-increment the counter
let v4 = db.volatile(); // second call will be cached
@@ -40,7 +40,7 @@ fn intermingled() {
assert_eq!(v1, v3);
assert_eq!(v2, v4);
- db.salsa_runtime_mut().synthetic_write(Durability::LOW); // clears volatile caches
+ db.synthetic_write(Durability::LOW); // clears volatile caches
let v5 = db.memoized(); // re-executes volatile, caches new result
let v6 = db.memoized(); // re-use cached result
diff --git a/crates/stdx/src/lib.rs b/crates/stdx/src/lib.rs
index 9a9ebae74e..0504ca50b8 100644
--- a/crates/stdx/src/lib.rs
+++ b/crates/stdx/src/lib.rs
@@ -302,6 +302,22 @@ pub fn slice_tails<T>(this: &[T]) -> impl Iterator<Item = &[T]> {
(0..this.len()).map(|i| &this[i..])
}
+pub trait IsNoneOr {
+ type Type;
+ #[allow(clippy::wrong_self_convention)]
+ fn is_none_or(self, s: impl FnOnce(Self::Type) -> bool) -> bool;
+}
+#[allow(unstable_name_collisions)]
+impl<T> IsNoneOr for Option<T> {
+ type Type = T;
+ fn is_none_or(self, f: impl FnOnce(T) -> bool) -> bool {
+ match self {
+ Some(v) => f(v),
+ None => true,
+ }
+ }
+}
+
#[cfg(test)]
mod tests {
use super::*;
diff --git a/editors/code/src/commands.ts b/editors/code/src/commands.ts
index 3d33d255ad..849fae5cf2 100644
--- a/editors/code/src/commands.ts
+++ b/editors/code/src/commands.ts
@@ -4,7 +4,11 @@ import * as ra from "./lsp_ext";
import * as path from "path";
import type { Ctx, Cmd, CtxInit } from "./ctx";
-import { applySnippetWorkspaceEdit, applySnippetTextEdits } from "./snippets";
+import {
+ applySnippetWorkspaceEdit,
+ applySnippetTextEdits,
+ type SnippetTextDocumentEdit,
+} from "./snippets";
import { spawnSync } from "child_process";
import { type RunnableQuickPick, selectRunnable, createTask, createArgs } from "./run";
import { AstInspector } from "./ast_inspector";
@@ -1006,7 +1010,6 @@ export function resolveCodeAction(ctx: CtxInit): Cmd {
return;
}
const itemEdit = item.edit;
- const edit = await client.protocol2CodeConverter.asWorkspaceEdit(itemEdit);
// filter out all text edits and recreate the WorkspaceEdit without them so we can apply
// snippet edits on our own
const lcFileSystemEdit = {
@@ -1017,16 +1020,71 @@ export function resolveCodeAction(ctx: CtxInit): Cmd {
lcFileSystemEdit,
);
await vscode.workspace.applyEdit(fileSystemEdit);
- await applySnippetWorkspaceEdit(edit);
+
+ // replace all text edits so that we can convert snippet text edits into `vscode.SnippetTextEdit`s
+ // FIXME: this is a workaround until vscode-languageclient supports doing the SnippeTextEdit conversion itself
+ // also need to carry the snippetTextDocumentEdits separately, since we can't retrieve them again using WorkspaceEdit.entries
+ const [workspaceTextEdit, snippetTextDocumentEdits] = asWorkspaceSnippetEdit(ctx, itemEdit);
+ await applySnippetWorkspaceEdit(workspaceTextEdit, snippetTextDocumentEdits);
if (item.command != null) {
await vscode.commands.executeCommand(item.command.command, item.command.arguments);
}
};
}
+function asWorkspaceSnippetEdit(
+ ctx: CtxInit,
+ item: lc.WorkspaceEdit,
+): [vscode.WorkspaceEdit, SnippetTextDocumentEdit[]] {
+ const client = ctx.client;
+
+ // partially borrowed from https://github.com/microsoft/vscode-languageserver-node/blob/295aaa393fda8ecce110c38880a00466b9320e63/client/src/common/protocolConverter.ts#L1060-L1101
+ const result = new vscode.WorkspaceEdit();
+
+ if (item.documentChanges) {
+ const snippetTextDocumentEdits: SnippetTextDocumentEdit[] = [];
+
+ for (const change of item.documentChanges) {
+ if (lc.TextDocumentEdit.is(change)) {
+ const uri = client.protocol2CodeConverter.asUri(change.textDocument.uri);
+ const snippetTextEdits: (vscode.TextEdit | vscode.SnippetTextEdit)[] = [];
+
+ for (const edit of change.edits) {
+ if (
+ "insertTextFormat" in edit &&
+ edit.insertTextFormat === lc.InsertTextFormat.Snippet
+ ) {
+ // is a snippet text edit
+ snippetTextEdits.push(
+ new vscode.SnippetTextEdit(
+ client.protocol2CodeConverter.asRange(edit.range),
+ new vscode.SnippetString(edit.newText),
+ ),
+ );
+ } else {
+ // always as a text document edit
+ snippetTextEdits.push(
+ vscode.TextEdit.replace(
+ client.protocol2CodeConverter.asRange(edit.range),
+ edit.newText,
+ ),
+ );
+ }
+ }
+
+ snippetTextDocumentEdits.push([uri, snippetTextEdits]);
+ }
+ }
+ return [result, snippetTextDocumentEdits];
+ } else {
+ // we don't handle WorkspaceEdit.changes since it's not relevant for code actions
+ return [result, []];
+ }
+}
+
export function applySnippetWorkspaceEditCommand(_ctx: CtxInit): Cmd {
return async (edit: vscode.WorkspaceEdit) => {
- await applySnippetWorkspaceEdit(edit);
+ await applySnippetWorkspaceEdit(edit, edit.entries());
};
}
diff --git a/editors/code/src/snippets.ts b/editors/code/src/snippets.ts
index d81765649f..b3982bdf2b 100644
--- a/editors/code/src/snippets.ts
+++ b/editors/code/src/snippets.ts
@@ -3,20 +3,28 @@ import * as vscode from "vscode";
import { assert } from "./util";
import { unwrapUndefinable } from "./undefinable";
-export async function applySnippetWorkspaceEdit(edit: vscode.WorkspaceEdit) {
- if (edit.entries().length === 1) {
- const [uri, edits] = unwrapUndefinable(edit.entries()[0]);
+export type SnippetTextDocumentEdit = [vscode.Uri, (vscode.TextEdit | vscode.SnippetTextEdit)[]];
+
+export async function applySnippetWorkspaceEdit(
+ edit: vscode.WorkspaceEdit,
+ editEntries: SnippetTextDocumentEdit[],
+) {
+ if (editEntries.length === 1) {
+ const [uri, edits] = unwrapUndefinable(editEntries[0]);
const editor = await editorFromUri(uri);
- if (editor) await applySnippetTextEdits(editor, edits);
+ if (editor) {
+ edit.set(uri, removeLeadingWhitespace(editor, edits));
+ await vscode.workspace.applyEdit(edit);
+ }
return;
}
- for (const [uri, edits] of edit.entries()) {
+ for (const [uri, edits] of editEntries) {
const editor = await editorFromUri(uri);
if (editor) {
await editor.edit((builder) => {
for (const indel of edits) {
assert(
- !parseSnippet(indel.newText),
+ !(indel instanceof vscode.SnippetTextEdit),
`bad ws edit: snippet received with multiple edits: ${JSON.stringify(
edit,
)}`,
@@ -39,53 +47,97 @@ async function editorFromUri(uri: vscode.Uri): Promise<vscode.TextEditor | undef
}
export async function applySnippetTextEdits(editor: vscode.TextEditor, edits: vscode.TextEdit[]) {
- const selections: vscode.Selection[] = [];
- let lineDelta = 0;
- await editor.edit((builder) => {
- for (const indel of edits) {
- const parsed = parseSnippet(indel.newText);
- if (parsed) {
- const [newText, [placeholderStart, placeholderLength]] = parsed;
- const prefix = newText.substr(0, placeholderStart);
- const lastNewline = prefix.lastIndexOf("\n");
+ const edit = new vscode.WorkspaceEdit();
+ const snippetEdits = toSnippetTextEdits(edits);
+ edit.set(editor.document.uri, removeLeadingWhitespace(editor, snippetEdits));
+ await vscode.workspace.applyEdit(edit);
+}
- const startLine = indel.range.start.line + lineDelta + countLines(prefix);
- const startColumn =
- lastNewline === -1
- ? indel.range.start.character + placeholderStart
- : prefix.length - lastNewline - 1;
- const endColumn = startColumn + placeholderLength;
- selections.push(
- new vscode.Selection(
- new vscode.Position(startLine, startColumn),
- new vscode.Position(startLine, endColumn),
- ),
+function hasSnippet(snip: string): boolean {
+ const m = snip.match(/\$\d+|\{\d+:[^}]*\}/);
+ return m != null;
+}
+
+function toSnippetTextEdits(
+ edits: vscode.TextEdit[],
+): (vscode.TextEdit | vscode.SnippetTextEdit)[] {
+ return edits.map((textEdit) => {
+ // Note: text edits without any snippets are returned as-is instead of
+ // being wrapped in a SnippetTextEdit, as otherwise it would be
+ // treated as if it had a tab stop at the end.
+ if (hasSnippet(textEdit.newText)) {
+ return new vscode.SnippetTextEdit(
+ textEdit.range,
+ new vscode.SnippetString(textEdit.newText),
+ );
+ } else {
+ return textEdit;
+ }
+ });
+}
+
+/**
+ * Removes the leading whitespace from snippet edits, so as to not double up
+ * on indentation.
+ *
+ * Snippet edits by default adjust any multi-line snippets to match the
+ * indentation of the line to insert at. Unfortunately, we (the server) also
+ * include the required indentation to match what we line insert at, so we end
+ * up doubling up the indentation. Since there isn't any way to tell vscode to
+ * not fixup indentation for us, we instead opt to remove the indentation and
+ * then let vscode add it back in.
+ *
+ * This assumes that the source snippet text edits have the required
+ * indentation, but that's okay as even without this workaround and the problem
+ * to workaround, those snippet edits would already be inserting at the wrong
+ * indentation.
+ */
+function removeLeadingWhitespace(
+ editor: vscode.TextEditor,
+ edits: (vscode.TextEdit | vscode.SnippetTextEdit)[],
+) {
+ return edits.map((edit) => {
+ if (edit instanceof vscode.SnippetTextEdit) {
+ const snippetEdit: vscode.SnippetTextEdit = edit;
+ const firstLineEnd = snippetEdit.snippet.value.indexOf("\n");
+
+ if (firstLineEnd !== -1) {
+ // Is a multi-line snippet, remove the indentation which
+ // would be added back in by vscode.
+ const startLine = editor.document.lineAt(snippetEdit.range.start.line);
+ const leadingWhitespace = getLeadingWhitespace(
+ startLine.text,
+ 0,
+ startLine.firstNonWhitespaceCharacterIndex,
);
- builder.replace(indel.range, newText);
- } else {
- builder.replace(indel.range, indel.newText);
+
+ const [firstLine, rest] = splitAt(snippetEdit.snippet.value, firstLineEnd + 1);
+ const unindentedLines = rest
+ .split("\n")
+ .map((line) => line.replace(leadingWhitespace, ""))
+ .join("\n");
+
+ snippetEdit.snippet.value = firstLine + unindentedLines;
}
- lineDelta +=
- countLines(indel.newText) - (indel.range.end.line - indel.range.start.line);
+
+ return snippetEdit;
+ } else {
+ return edit;
}
});
- if (selections.length > 0) editor.selections = selections;
- if (selections.length === 1) {
- const selection = unwrapUndefinable(selections[0]);
- editor.revealRange(selection, vscode.TextEditorRevealType.InCenterIfOutsideViewport);
- }
}
-function parseSnippet(snip: string): [string, [number, number]] | undefined {
- const m = snip.match(/\$(0|\{0:([^}]*)\})/);
- if (!m) return undefined;
- const placeholder = m[2] ?? "";
- if (m.index == null) return undefined;
- const range: [number, number] = [m.index, placeholder.length];
- const insert = snip.replace(m[0], placeholder);
- return [insert, range];
+// based on https://github.com/microsoft/vscode/blob/main/src/vs/base/common/strings.ts#L284
+function getLeadingWhitespace(str: string, start: number = 0, end: number = str.length): string {
+ for (let i = start; i < end; i++) {
+ const chCode = str.charCodeAt(i);
+ if (chCode !== " ".charCodeAt(0) && chCode !== " ".charCodeAt(0)) {
+ return str.substring(start, i);
+ }
+ }
+ return str.substring(start, end);
}
-function countLines(text: string): number {
- return (text.match(/\n/g) || []).length;
+function splitAt(str: string, index: number): [string, string] {
+ return [str.substring(0, index), str.substring(index)];
}
diff --git a/xtask/src/metrics.rs b/xtask/src/metrics.rs
index 2efafa10a8..285abb9efc 100644
--- a/xtask/src/metrics.rs
+++ b/xtask/src/metrics.rs
@@ -86,7 +86,11 @@ impl Metrics {
fn measure_rustc_tests(&mut self, sh: &Shell) -> anyhow::Result<()> {
eprintln!("\nMeasuring rustc tests");
- cmd!(sh, "git clone --depth=1 https://github.com/rust-lang/rust").run()?;
+ cmd!(
+ sh,
+ "git clone --depth=1 --branch 1.76.0 https://github.com/rust-lang/rust.git --single-branch"
+ )
+ .run()?;
let output = cmd!(sh, "./target/release/rust-analyzer rustc-tests ./rust").read()?;
for (metric, value, unit) in parse_metrics(&output) {