I don't mean to complain. Doing software engineering for a living is a situation of extreme privilege. But there's something to be said about how alienating it can be at times.

Once, just once, I want to be able to answer someone's "what are you working on?" question with "see that house? it wasn't there last year. I built that".

Instead for now, I have to answer with: "well you see... support for proc macros was broken in rust-analyzer for folks who used a nightly rustc toolchain, due to incompatibilities in the bridge (which is an unstable interface in the first place), and it's bound to stay broken for the foreseeable future, not specifically because of technical challenges, but mostly because of human and organizational challenges, and I think I've found a way forward that will benefit everyone."

And that's not a good answer?

Not really, no. At that point, the asker is usually nervously looking at the time, and they start to blurt out some excuse to cut the conversation short, just as I was about to ask them if they'd ever heard of Conway's law.

Luckily, blogging about it is a completely different proposition. I have a bit of a reputation. You could stop reading at any time. Now would be a good time. You could stop now, too! Nope? Still here? Alright then.

Let's dig in.

A tale of one and a half compilers

There's always a lot of worry that Rust is defined by its "only implementation", rustc. But that's not true!

First of all, Ferrous Systems just announced The Ferrocene Language Specification, which I hear we can use to silence arguments now?

Oh you.

Anyway it's still a work-in-progress and it definitely doesn't normate (?) the version rustc I use daily. But it's a thing. mrustc is also a thing! All I know about it is that it's written in C++, for the only purpose of "bootstrapping" Rust from C: you build (an older) rustc with it, and then you build a newer rustc, and finally you have a working, recent rustc.

Compilers are cool!

And then there's gccrs, which elicited some... mixed reactions to say the least (but I'm reserving my judgement until later - much, much later). Not to be confused with rustc_codegen_gcc, which is "just" a codegen backend for rustc, the One True Compiler.

But those aren't overly relevant to today's topic.

The other compiler I want to talk about is... rust-analyzer!

Wait. rust-analyzer, a compiler?

Isn't it just there to provide code intelligence, and assists?

Why yes! And how do you suppose it does that?

Well it... it uhh... mh.

Well! It uses rustc's lexer, but has its own parser.

It has its own AST types, which leverage rowan. It uses salsa for on-demand, incremental queries (like "show me all references for this symbol"), whereas rustc uses its own, pre-salsa query system.

rust-analyzer uses chalk as a trait engine, whereas in rustc, chalk is available behind a flag, but it's far from being the default.

So yeah! I didn't know that a month ago, but: rust-analyzer really does re-implement a lot of things rustc does, in a slightly different way. And we could spend time talking about rustc's librarification, but we're not going to.

Instead, we'll just let those decision inform our impression of rust-analyzer as a separate project, with its own rules, its own design constraints, its own team, its own release schedule, and its own wild success β€” at the time of this writing, the VS Code extension alone has been downloaded almost a million times.

Two kinds of macros

rust-analyzer even has its own "macros by example" engine.

When you have some code like this:

Rust code
macro_rules! do_thrice {
		// We expect a block
    ($body: block) => {
				// And simply repeat it three times (or "thrice")
        $body
        $body
        $body
    };
}

fn main() {
    fn say_hi() {
        println!("Hi!");
    }

		// I originally had a `println!` directly in there, but since that's
		// a macro too, it also got expanded, making the example more confusing
		// than it needed to be.
    do_thrice! {{ say_hi() }}
}

And you position your cursor where the do_thrice macro is invoked, and pick the "Rust Analyzer: Expand macro recursively" command from the VS Code Command Palette, it gives you this (in a new tab):

Rust code
// Recursive expansion of do_thrice! macro
// ========================================

{
    say_hi()
}
{
    say_hi()
}
{
    say_hi()
}

...and that doesn't use any rustc machinery! It's all rust-analyzer doing its business on its own.

However... there's another kind of macro: procedural macros, or proc macros for short. And those are another story altogether.

Let's try and rewrite the do_thrice! macro as a proc macro. Not because it's a good idea (far from it), but because it'll make a fine introduction.

First we'll need to make a separate crate:

Shell session
$ cargo new --lib pm
     Created library `pm` package

And we'll adjust its Cargo.toml so it becomes a proc-macro crate:

TOML markup
[package]
name = "pm"
version = "0.1.0"
edition = "2021"

# πŸ‘‡ here's the important bit
[lib]
proc-macro = true

[dependencies]

And in src/lib.rs, we'll have:

Rust code
use proc_macro::TokenStream;

#[proc_macro]
pub fn do_thrice(args: TokenStream) -> TokenStream {
    let mut stream = TokenStream::default();
    stream.extend(args.clone());
    stream.extend(args.clone());
    stream.extend(args);
    stream
}

Oh, we're manipulating TokenStreams directly from Rust code!

Yes, that's the "procedural" in "proc macros".

And then let's call it from some other crate:

Shell session
$ cd sample/
$ cargo add --path ../pm
      Adding pm (local) to dependencies.
Rust code
// in `sample/src/main.rs`

fn main() {
    fn say_hi() {
        println!("Hi!");
    }

    pm::do_thrice! {{ say_hi(); }}
}

And then that works just the same:

Shell session
$ cargo run
   Compiling sample v0.1.0 (/home/amos/bearcove/sample)
    Finished dev [unoptimized + debuginfo] target(s) in 0.22s
     Running `target/debug/sample`
Hi!
Hi!
Hi!

But let's take a closer look at what happens when we compile our sample executable. No! When we check it:

Shell session
$ cargo clean --quiet && cargo check
   Compiling pm v0.1.0 (/home/amos/bearcove/pm)
    Checking sample v0.1.0 (/home/amos/bearcove/sample)
    Finished dev [unoptimized + debuginfo] target(s) in 0.40s

Mhh.. it says "Checking" for sample, but it says "Compiling" for pm.

Yup! In fact, let's go one step further and run it without incremental compilation (this'll reduce the number of files we need to look at):

Shell session
$ cargo clean --quiet && CARGO_INCREMENTAL=0 cargo check
   Compiling pm v0.1.0 (/home/amos/bearcove/pm)
    Checking sample v0.1.0 (/home/amos/bearcove/sample)
    Finished dev [unoptimized + debuginfo] target(s) in 0.35s

...and look inside target/:

Shell session
$ tree -ah target                                       
[4.0K]  target
β”œβ”€β”€ [ 177]  CACHEDIR.TAG
β”œβ”€β”€ [4.0K]  debug
β”‚   β”œβ”€β”€ [4.0K]  build
β”‚   β”œβ”€β”€ [   0]  .cargo-lock
β”‚   β”œβ”€β”€ [4.0K]  deps
β”‚   β”‚   β”œβ”€β”€ [7.7M]  libpm-29e65e9cc9cd67f3.so
β”‚   β”‚   β”œβ”€β”€ [   0]  libsample-1e75863094962883.rmeta
β”‚   β”‚   β”œβ”€β”€ [ 245]  pm-29e65e9cc9cd67f3.d
β”‚   β”‚   └── [ 187]  sample-1e75863094962883.d
β”‚   β”œβ”€β”€ [4.0K]  examples
β”‚   β”œβ”€β”€ [4.0K]  .fingerprint
β”‚   β”‚   β”œβ”€β”€ [4.0K]  pm-29e65e9cc9cd67f3
β”‚   β”‚   β”‚   β”œβ”€β”€ [  23]  dep-lib-pm
β”‚   β”‚   β”‚   β”œβ”€β”€ [  48]  invoked.timestamp
β”‚   β”‚   β”‚   β”œβ”€β”€ [  16]  lib-pm
β”‚   β”‚   β”‚   └── [ 328]  lib-pm.json
β”‚   β”‚   └── [4.0K]  sample-1e75863094962883
β”‚   β”‚       β”œβ”€β”€ [  16]  bin-sample
β”‚   β”‚       β”œβ”€β”€ [ 388]  bin-sample.json
β”‚   β”‚       β”œβ”€β”€ [  24]  dep-bin-sample
β”‚   β”‚       └── [  48]  invoked.timestamp
β”‚   └── [4.0K]  incremental
└── [1.1K]  .rustc_info.json

8 directories, 15 files

We can see in target/debug/deps, a libpm-HASH.so file. So pm was compiled, and linked. But we can't see one for sample! Just .rmeta and .d files.

Well, maybe it compiles all dependencies, and "checks" only our crate?

Ah, let's try! I found a crate with zero external dependencies: lhash.

Let's add it and use it:

Shell session
$ cargo add lhash -F sha512
    Updating crates.io index
      Adding lhash v1.0.1 to dependencies.
             Features:
             + sha512
             - md5
             - sha1
             - sha256

This is not an endorsement of the lhash crate. I have not reviewed its correctness or security. I just know it has zero dependencies, so it's convenient for our little experiment.

Rust code
use lhash::sha512;

fn main() {
    fn say_hi() {
        println!("Hi! sha512(\"hi\") = {:x?}", sha512(b"hi"));
    }

    pm::do_thrice! {{ say_hi(); }}
}
Shell session
$ cargo clean --quiet && CARGO_INCREMENTAL=0 cargo check
    Checking lhash v1.0.1
   Compiling pm v0.1.0 (/home/amos/bearcove/pm)
    Checking sample v0.1.0 (/home/amos/bearcove/sample)
    Finished dev [unoptimized + debuginfo] target(s) in 0.35s

No, see! It's only checking lhash, not compiling it.

Shell session
$ tree -ah target/debug/deps
[4.0K]  target/debug/deps
β”œβ”€β”€ [ 651]  lhash-2f1fe434799a22fd.d
β”œβ”€β”€ [133K]  liblhash-2f1fe434799a22fd.rmeta
β”œβ”€β”€ [7.7M]  libpm-29e65e9cc9cd67f3.so
β”œβ”€β”€ [   0]  libsample-6bf261eaf0ec4223.rmeta
β”œβ”€β”€ [ 245]  pm-29e65e9cc9cd67f3.d
└── [ 187]  sample-6bf261eaf0ec4223.d

0 directories, 6 files

Whereas if we actually build our crate:

Shell session
$ cargo clean --quiet && CARGO_INCREMENTAL=0 cargo build
   Compiling pm v0.1.0 (/home/amos/bearcove/pm)
   Compiling lhash v1.0.1
   Compiling sample v0.1.0 (/home/amos/bearcove/sample)
    Finished dev [unoptimized + debuginfo] target(s) in 0.46s

All crates are actually compiled:

Shell session
$ tree -ah target/debug/deps                            
[4.0K]  target/debug/deps
β”œβ”€β”€ [ 896]  lhash-83b88ad4daff567f.d
β”œβ”€β”€ [336K]  liblhash-83b88ad4daff567f.rlib
β”œβ”€β”€ [157K]  liblhash-83b88ad4daff567f.rmeta
β”œβ”€β”€ [7.7M]  libpm-29e65e9cc9cd67f3.so
β”œβ”€β”€ [ 245]  pm-29e65e9cc9cd67f3.d
β”œβ”€β”€ [3.8M]  sample-603096a51bd3460d
└── [ 181]  sample-603096a51bd3460d.d

0 directories, 7 files

We've got our sample executable:

Shell session
$ ./target/debug/deps/sample-603096a51bd3460d 
Hi! sha512("hi") = [15, a, 14, ed, 5b, ea, 6c, c7, 31, cf, 86, c4, 15, 66, ac, 42, 7a, 8d, b4, 8e, f1, b9, fd, 62, 66, 64, b3, bf, bb, 99, 7, 1f, a4, c9, 22, f3, 3d, de, 38, 71, 9b, 8c, 83, 54, e2, b7, ab, 9d, 77, e0, e6, 7f, c1, 28, 43, 92, a, 71, 2e, 73, d5, 58, e1, 97]
Hi! sha512("hi") = [15, a, 14, ed, 5b, ea, 6c, c7, 31, cf, 86, c4, 15, 66, ac, 42, 7a, 8d, b4, 8e, f1, b9, fd, 62, 66, 64, b3, bf, bb, 99, 7, 1f, a4, c9, 22, f3, 3d, de, 38, 71, 9b, 8c, 83, 54, e2, b7, ab, 9d, 77, e0, e6, 7f, c1, 28, 43, 92, a, 71, 2e, 73, d5, 58, e1, 97]
Hi! sha512("hi") = [15, a, 14, ed, 5b, ea, 6c, c7, 31, cf, 86, c4, 15, 66, ac, 42, 7a, 8d, b4, 8e, f1, b9, fd, 62, 66, 64, b3, bf, bb, 99, 7, 1f, a4, c9, 22, f3, 3d, de, 38, 71, 9b, 8c, 83, 54, e2, b7, ab, 9d, 77, e0, e6, 7f, c1, 28, 43, 92, a, 71, 2e, 73, d5, 58, e1, 97]

lhash became an .rlib, which, at least on this platform is an "ar archive": it exports symbols.

Shell session
$ llvm-nm -C target/debug/deps/liblhash-83b88ad4daff567f.rlib | grep sha512
target/debug/deps/liblhash-83b88ad4daff567f.rlib:lib.rmeta: no symbols
0000000000000000 T lhash::sha512::sha512_transform::h9caf21c0d8b3b830
0000000000000000 T lhash::sha512::sha512_transform::read_u64::h676973d5422c7cb3
0000000000000000 T lhash::sha512::Sha512::const_result::hc60c40db03c4dd8b
0000000000000000 T lhash::sha512::Sha512::const_update::hbb8f86fdc2ac7014
0000000000000000 T lhash::sha512::Sha512::new::h8c0606f3ad6bec45
0000000000000000 T lhash::sha512::Sha512::reset::ha635bfd5375d19ff
0000000000000000 T lhash::sha512::Sha512::result::hdbd28d30df99667a
0000000000000000 T lhash::sha512::Sha512::update::h4612ff723705776

As for pm, our proc-macro crate, well... it's an .so, which on Linux is a shared object (called a dynamic library / dylib on other platforms). It also exports symbols:

Shell session
$ llvm-nm -C target/debug/deps/libpm-29e65e9cc9cd67f3.so | grep proc_macro | head
000000000006f290 t _$LT$proc_macro..bridge..client..TokenStream$u20$as$u20$proc_macro..bridge..rpc..Encode$LT$S$GT$$GT$::encode::h0cedb28ac4dc5a9d
000000000006f2f0 t _$LT$proc_macro..bridge..client..TokenStream$u20$as$u20$proc_macro..bridge..rpc..DecodeMut$LT$S$GT$$GT$::decode::h9ed6799f031aa8a9
0000000000074290 t _$LT$proc_macro..TokenStream$u20$as$u20$core..iter..traits..collect..Extend$LT$proc_macro..TokenTree$GT$$GT$::extend::h4960c6a4e1c9de92
0000000000095530 T proc_macro::SourceFile::path::hd8caf5f9f645aac9
00000000000959e0 T proc_macro::SourceFile::is_real::h8aa09afecc5cf1c5
0000000000080720 t proc_macro::diagnostic::Diagnostic::emit::to_internal::hb72589cd22aa2875
000000000007edd0 T proc_macro::diagnostic::Diagnostic::emit::h906bdda238951179
000000000007ed70 T proc_macro::diagnostic::Diagnostic::level::hd25e9fc78c03917e
000000000007eda0 T proc_macro::diagnostic::Diagnostic::spans::hd804305810893e73
000000000007ed90 T proc_macro::diagnostic::Diagnostic::message::h872f0bf8b64b7c5d

And all of that makes sense... if you consider how proc macros must work.

Proc macros in rustc

The most important symbol exported by the proc-macro's shared object / dynamic library is this one:

Shell session
$ llvm-nm -C target/debug/deps/libpm-29e65e9cc9cd67f3.so | grep rustc_proc_macro_decls
00000000001b55e8 D __rustc_proc_macro_decls_e4a0d4906233add5__

It's a symbol of type &&[proc_macro::bridge::client::ProcMacro], and that ProcMacro type is an enum, like so:

Rust code
#[repr(C)]
#[derive(Copy, Clone)]
pub enum ProcMacro {
    CustomDerive {
        trait_name: &'static str,
        attributes: &'static [&'static str],
        client: Client<fn(crate::TokenStream) -> crate::TokenStream>,
    },

    Attr {
        name: &'static str,
        client: Client<fn(crate::TokenStream, crate::TokenStream) -> crate::TokenStream>,
    },

    Bang {
        name: &'static str,
        client: Client<fn(crate::TokenStream) -> crate::TokenStream>,
    },
}

Wait, #[repr(C)] on a data-carrying enum?

Hey look at you using proper terminology, have you been doing some reading? If so, you must know that RFC 2195 defines the layout of #[repr(C)] enums, even non-C-like enums.

In other words: it's the proper representation to use if you're planning on doing FFI (foreign function interfacing), which is very much what's happening here.

So, a compiler like rustc would:

And from there, if it needs to expand a macro, it just needs to find it by name (by matching against the trait_name / name), and then it uses the client, of type Client:

Rust code
#[repr(C)]
#[derive(Copy, Clone)]
pub struct Client<F> {
    // FIXME(eddyb) use a reference to the `static COUNTERS`, instead of
    // a wrapper `fn` pointer, once `const fn` can reference `static`s.
    pub(super) get_handle_counters: extern "C" fn() -> &'static HandleCounters,
    pub(super) run: extern "C" fn(Bridge<'_>, F) -> Buffer<u8>,
    pub(super) f: F,
}

Again, notice how ffi-safe all of this is - repr(C), extern "C", only function pointers, no Fn, etc.

Even the Buffer type there is repr(C) and contains raw pointers + a vtable:

Rust code
#[repr(C)]
pub struct Buffer<T: Copy> {
    data: *mut T,
    len: usize,
    capacity: usize,
    reserve: extern "C" fn(Buffer<T>, usize) -> Buffer<T>,
    drop: extern "C" fn(Buffer<T>),
}

Anyway in Client, there's a run function, and you can call it via the convenience Client::run method:

Rust code
impl client::Client<fn(crate::TokenStream, crate::TokenStream) -> crate::TokenStream> {
    pub fn run<S: Server>(
        &self,
        strategy: &impl ExecutionStrategy,
        server: S,
        input: S::TokenStream,
        input2: S::TokenStream,
        force_show_panics: bool,
    ) -> Result<S::TokenStream, PanicMessage> {
        let client::Client { get_handle_counters, run, f } = *self;
        run_server(
            strategy,
            get_handle_counters(),
            server,
            (
                <MarkedTypes<S> as Types>::TokenStream::mark(input),
                <MarkedTypes<S> as Types>::TokenStream::mark(input2),
            ),
            run,
            f,
            force_show_panics,
        )
        .map(<MarkedTypes<S> as Types>::TokenStream::unmark)
    }
}

Which... takes a server.

Wait, so the proc macro dylib is the client?

Yeah! It goes like this:

All the proc-macro dylib does is manipulate data structures (token streams, token trees, literals, groups, etc.). But what's interesting is that these are all defined by the server! It can use any type it wants!

In fact, one of the traits a server must implement is Types, and it's just associated types:

Rust code
// in `rust/compiler/rustc_expand/src/proc_macro_server.rs`

impl server::Types for Rustc<'_, '_> {
    type FreeFunctions = FreeFunctions;
    type TokenStream = TokenStream;
    type SourceFile = Lrc<SourceFile>;
    type MultiSpan = Vec<Span>;
    type Diagnostic = Diagnostic;
    type Span = Span;
    type Symbol = Symbol;
}

And then there's separate traits like TokenStream, for example - here's just a couple of its methods:

Rust code
// in `rust/compiler/rustc_expand/src/proc_macro_server.rs`

impl server::TokenStream for Rustc<'_, '_> {
    fn is_empty(&mut self, stream: &Self::TokenStream) -> bool {
        stream.is_empty()
    }

    fn from_str(&mut self, src: &str) -> Self::TokenStream {
        parse_stream_from_source_str(
            FileName::proc_macro_source_code(src),
            src.to_string(),
            self.sess(),
            Some(self.call_site),
        )
    }

    fn to_string(&mut self, stream: &Self::TokenStream) -> String {
        pprust::tts_to_string(stream)
    }

    // etc.
}

Oooh right - we're using the server's TokenStream type because it's the server that lexes source code in the first place!

Yeah! It's the rust compiler that turns code like foo(bar) into something like [Ident("foo"), Group(Parenthesis, Ident("bar"))], and the proc macro is merely calling methods to manipulate that stream.

Right, right right right.

And identifiers, for example, are "interned" by the proc-macro server. If you have the identifier "foo" a hundred times, you don't pass around a "foo" string every time. The first time it's encountered, it's added to an interner (you can think of it as a gigantic array) and after that, you pass around an index, which is just a small, nice, Copy integer type.

Okay so let me recap. The proc macro server (rustc) parses all the source files, and when it encounters a proc macros, it loads the appropriate dylib, finds the registrar symbol, finds the relevant proc macro, "runs it" acting as a server, the proc macro manipulates types owned by the server and... then that's it?

Yeah! rustc gets the code "generated" or transformed by the proc macro, and it can continue compiling stuff (which involves actually parsing it from tokens into an AST, then various intermediate representations, then most often LLVM IR, then whatever binary format you're targetting).

I see. And then... for rust-analyzer...

Well, here's where it gets fun.

Proc macros in rust-analyzer

We've already established that rust-analyzer is half a compiler - it's a frontend with some code analysis features added on top.

So it has a different TokenStream type than rustc, for example. It has a different strategy for interning symbols, it thinks about spans differently: because it was developed in isolation, there's not much that looks alike except for some high-level principles (and a best-effort shot at compatibility).

And it mostly works for "macros by example", because it "just" reimplements the substitution logic. But proc macros cannot be expanded in isolation. You have to compile the proc macro crate to a library first: you may have seen a "cannot expand proc macro because crate hasn't been compiled yet" error message for example. That's why!

Luckily rust-analyzer already needs to run cargo check for other reasons (because most diagnostics come from it, rather than rust-analyzer itself). That has the nice side-effect of compiling all proc macros to dylibs, as we've seen.

And then after that, rust-analyzer has to act as a proc macro server.

The only other mainstream proc macro server in existence, after rustc.

And so rust-analyzer also implements all the same traits, Types for example:

Rust code
// in `rust-analyzer/crates/proc-macro-srv/src/abis/abi_sysroot/ra_server.rs`

impl server::Types for RustAnalyzer {
    type FreeFunctions = FreeFunctions;
    type TokenStream = TokenStream;
    type SourceFile = SourceFile;
    type MultiSpan = Vec<Span>;
    type Diagnostic = Diagnostic;
    type Span = Span;
    type Symbol = Symbol;
}

But also TokenStream:

Rust code
// in `rust-analyzer/crates/proc-macro-srv/src/abis/abi_sysroot/ra_server.rs`

impl server::TokenStream for RustAnalyzer {
    fn is_empty(&mut self, stream: &Self::TokenStream) -> bool {
        stream.is_empty()
    }
    fn from_str(&mut self, src: &str) -> Self::TokenStream {
        use std::str::FromStr;

        Self::TokenStream::from_str(src).expect("cannot parse string")
    }
    fn to_string(&mut self, stream: &Self::TokenStream) -> String {
        stream.to_string()
    }

    // etc.
}

And everything else. And just like rustc, when it needs to expand a proc macro, it... assumes cargo check was run earlier on by some other parts of rust-analyzer, finds the proc-macro dylib, opens it (with RTLD_DEEPBIND on Linux because of this fun thing), finds the registrar symbol, enumerates all available proc macros, and runs the one it needs, then carries on indexing the generated code so it can provide code intelligence just like it does for any other code.

Its proc macro server was originally introduced in April 2020, as a standalone executable (a bin package). Later on, it was moved to a subcommand of the main rust-analyzer executable, so that there's only one executable to ship.

The main rust-analyzer process and the proc-macro server communicate over a JSON interface, and so, we can do this:

Shell session
$ echo '{"ListMacros":{"dylib_path":"target/debug/deps/libpm-e886d9f9eaf24619.so"}}' | ~/.cargo/bin/rust-analyzer proc-macro
{"ListMacros":{"Ok":[["do_thrice","FuncLike"]]}}

Whoaaaa, neat!

Right? And there's also an ExpandMacro request type, which is a little more complicated to invoke by hand:

Rust code
// in `rust-analyzer/crates/proc-macro-api/src/msg.rs`

#[derive(Debug, Serialize, Deserialize)]
pub struct ExpandMacro {
    /// Argument of macro call.
    ///
    /// In custom derive this will be a struct or enum; in attribute-like macro - underlying
    /// item; in function-like macro - the macro body.
    pub macro_body: FlatTree,

    /// Name of macro to expand.
    ///
    /// In custom derive this is the name of the derived trait (`Serialize`, `Getters`, etc.).
    /// In attribute-like and function-like macros - single name of macro itself (`show_streams`).
    pub macro_name: String,

    /// Possible attributes for the attribute-like macros.
    pub attributes: Option<FlatTree>,

    pub lib: PathBuf,

    /// Environment variables to set during macro expansion.
    pub env: Vec<(String, String)>,

    pub current_dir: Option<String>,
}

Where FlatTree is a u32-based format. For example, identifiers are stored as a pair of u32:

Rust code
// in `rust-analyzer/crates/proc-macro-api/src/msg/flat.rs`

struct IdentRepr {
    id: tt::TokenId,
    text: u32,
}

impl IdentRepr {
    fn write(self) -> [u32; 2] {
        [self.id.0, self.text]
    }
    fn read(data: [u32; 2]) -> IdentRepr {
        IdentRepr { id: TokenId(data[0]), text: data[1] }
    }
}

Where the text field refers to an interned string:

Rust code
// in `rust-analyzer/crates/proc-macro-api/src/msg/flat.rs`

impl<'a> Writer<'a> {
    pub(crate) fn intern(&mut self, text: &'a str) -> u32 {
        let table = &mut self.text;
        *self.string_table.entry(text).or_insert_with(|| {
            let idx = table.len();
            table.push(text.to_string());
            idx as u32
        })
    }

    // etc.
}

And the string table itself is simply a HashMap<&'a str, u32>.

This all looks fine, to me.

Oh, yes, yes, that's not the problematic part.

There's a problematic part?

There absolutely is. Or was. Well. Soon-was.

The proc macro bridge interface

So as I mentioned, both rustc and rust-analyzer act as "proc macro servers", implementing a bunch of traits so that proc macros can interact with them to manipulate token streams.

But I haven't really said where those traits came from.

And the answer is... the proc_macro crate! You won't find it on https://crates.io, because it's a sysroot crate - like std or core. In fact, there's a bunch of those:

$ ls -lh "$(rustc --print sysroot)/lib/rustlib/src/rust/library"
total 64K
drwxrwxr-x 5 amos amos 4.0K Jul  7 11:14 alloc
drwxrwxr-x 8 amos amos 4.0K Jul  7 11:14 backtrace
drwxrwxr-x 6 amos amos 4.0K Jul  7 11:14 core
drwxrwxr-x 3 amos amos 4.0K Jul  7 11:14 panic_abort
drwxrwxr-x 3 amos amos 4.0K Jul  7 11:14 panic_unwind
drwxrwxr-x 4 amos amos 4.0K Jul  7 11:14 portable-simd
drwxrwxr-x 3 amos amos 4.0K Jul  7 11:14 proc_macro
drwxrwxr-x 3 amos amos 4.0K Jul  7 11:14 profiler_builtins
drwxrwxr-x 2 amos amos 4.0K Jul  7 11:14 rtstartup
drwxrwxr-x 2 amos amos 4.0K Jul  7 11:14 rustc-std-workspace-alloc
drwxrwxr-x 2 amos amos 4.0K Jul  7 11:14 rustc-std-workspace-core
drwxrwxr-x 2 amos amos 4.0K Jul  7 11:14 rustc-std-workspace-std
drwxrwxr-x 6 amos amos 4.0K Jul  7 11:14 std
drwxrwxr-x 6 amos amos 4.0K Jul  7 11:14 stdarch
drwxrwxr-x 3 amos amos 4.0K Jul  7 11:14 test
drwxrwxr-x 3 amos amos 4.0K Jul  7 11:14 unwind

And if we look in there, we can find the traits I was talking about:

Rust code
// in `proc_macro/src/bridge/server.rs`

macro_rules! declare_server_traits {
    ($($name:ident {
        $(fn $method:ident($($arg:ident: $arg_ty:ty),* $(,)?) $(-> $ret_ty:ty)?;)*
    }),* $(,)?) => {
        pub trait Types {
            $(associated_item!(type $name);)*
        }

        $(pub trait $name: Types {
            $(associated_item!(fn $method(&mut self, $($arg: $arg_ty),*) $(-> $ret_ty)?);)*
        })*

        pub trait Server: Types $(+ $name)* {}
        impl<S: Types $(+ $name)*> Server for S {}
    }
}
with_api!(Self, self_, declare_server_traits);

In a non-ironic, just funny, twist of fate, these traits are declared using "macros by example". Thankfully they're not using proc macros, that would be a little too much dogfooding for my taste.

You can see the full interface here:

Rust code
// in `proc_macro/src/bridge/mod.rs`

macro_rules! with_api {
    ($S:ident, $self:ident, $m:ident) => {
        $m! {
            FreeFunctions {
                fn drop($self: $S::FreeFunctions);
                fn track_env_var(var: &str, value: Option<&str>);
                fn track_path(path: &str);
            },
            TokenStream {
                fn drop($self: $S::TokenStream);
                fn clone($self: &$S::TokenStream) -> $S::TokenStream;
                fn new() -> $S::TokenStream;
                fn is_empty($self: &$S::TokenStream) -> bool;
                fn expand_expr($self: &$S::TokenStream) -> Result<$S::TokenStream, ()>;
                fn from_str(src: &str) -> $S::TokenStream;
                fn to_string($self: &$S::TokenStream) -> String;
                fn from_token_tree(
                    tree: TokenTree<$S::Group, $S::Punct, $S::Ident, $S::Literal>,
                ) -> $S::TokenStream;
                fn into_iter($self: $S::TokenStream) -> $S::TokenStreamIter;
            },
            TokenStreamBuilder {
                fn drop($self: $S::TokenStreamBuilder);
                fn new() -> $S::TokenStreamBuilder;
                fn push($self: &mut $S::TokenStreamBuilder, stream: $S::TokenStream);
                fn build($self: $S::TokenStreamBuilder) -> $S::TokenStream;
            },
            TokenStreamIter {
                fn drop($self: $S::TokenStreamIter);
                fn clone($self: &$S::TokenStreamIter) -> $S::TokenStreamIter;
                fn next(
                    $self: &mut $S::TokenStreamIter,
                ) -> Option<TokenTree<$S::Group, $S::Punct, $S::Ident, $S::Literal>>;
            },
            Group {
                fn drop($self: $S::Group);
                fn clone($self: &$S::Group) -> $S::Group;
                fn new(delimiter: Delimiter, stream: $S::TokenStream) -> $S::Group;
                fn delimiter($self: &$S::Group) -> Delimiter;
                fn stream($self: &$S::Group) -> $S::TokenStream;
                fn span($self: &$S::Group) -> $S::Span;
                fn span_open($self: &$S::Group) -> $S::Span;
                fn span_close($self: &$S::Group) -> $S::Span;
                fn set_span($self: &mut $S::Group, span: $S::Span);
            },
            Punct {
                fn new(ch: char, spacing: Spacing) -> $S::Punct;
                fn as_char($self: $S::Punct) -> char;
                fn spacing($self: $S::Punct) -> Spacing;
                fn span($self: $S::Punct) -> $S::Span;
                fn with_span($self: $S::Punct, span: $S::Span) -> $S::Punct;
            },
            Ident {
                fn new(string: &str, span: $S::Span, is_raw: bool) -> $S::Ident;
                fn span($self: $S::Ident) -> $S::Span;
                fn with_span($self: $S::Ident, span: $S::Span) -> $S::Ident;
            },
            Literal {
                fn drop($self: $S::Literal);
                fn clone($self: &$S::Literal) -> $S::Literal;
                fn from_str(s: &str) -> Result<$S::Literal, ()>;
                fn to_string($self: &$S::Literal) -> String;
                fn debug_kind($self: &$S::Literal) -> String;
                fn symbol($self: &$S::Literal) -> String;
                fn suffix($self: &$S::Literal) -> Option<String>;
                fn integer(n: &str) -> $S::Literal;
                fn typed_integer(n: &str, kind: &str) -> $S::Literal;
                fn float(n: &str) -> $S::Literal;
                fn f32(n: &str) -> $S::Literal;
                fn f64(n: &str) -> $S::Literal;
                fn string(string: &str) -> $S::Literal;
                fn character(ch: char) -> $S::Literal;
                fn byte_string(bytes: &[u8]) -> $S::Literal;
                fn span($self: &$S::Literal) -> $S::Span;
                fn set_span($self: &mut $S::Literal, span: $S::Span);
                fn subspan(
                    $self: &$S::Literal,
                    start: Bound<usize>,
                    end: Bound<usize>,
                ) -> Option<$S::Span>;
            },
            SourceFile {
                fn drop($self: $S::SourceFile);
                fn clone($self: &$S::SourceFile) -> $S::SourceFile;
                fn eq($self: &$S::SourceFile, other: &$S::SourceFile) -> bool;
                fn path($self: &$S::SourceFile) -> String;
                fn is_real($self: &$S::SourceFile) -> bool;
            },
            MultiSpan {
                fn drop($self: $S::MultiSpan);
                fn new() -> $S::MultiSpan;
                fn push($self: &mut $S::MultiSpan, span: $S::Span);
            },
            Diagnostic {
                fn drop($self: $S::Diagnostic);
                fn new(level: Level, msg: &str, span: $S::MultiSpan) -> $S::Diagnostic;
                fn sub(
                    $self: &mut $S::Diagnostic,
                    level: Level,
                    msg: &str,
                    span: $S::MultiSpan,
                );
                fn emit($self: $S::Diagnostic);
            },
            Span {
                fn debug($self: $S::Span) -> String;
                fn def_site() -> $S::Span;
                fn call_site() -> $S::Span;
                fn mixed_site() -> $S::Span;
                fn source_file($self: $S::Span) -> $S::SourceFile;
                fn parent($self: $S::Span) -> Option<$S::Span>;
                fn source($self: $S::Span) -> $S::Span;
                fn start($self: $S::Span) -> LineColumn;
                fn end($self: $S::Span) -> LineColumn;
                fn before($self: $S::Span) -> $S::Span;
                fn after($self: $S::Span) -> $S::Span;
                fn join($self: $S::Span, other: $S::Span) -> Option<$S::Span>;
                fn resolved_at($self: $S::Span, at: $S::Span) -> $S::Span;
                fn source_text($self: $S::Span) -> Option<String>;
                fn save_span($self: $S::Span) -> usize;
                fn recover_proc_macro_span(id: usize) -> $S::Span;
            },
        }
    };
}

The problem? All of this is extremely unstable.

Only a small part of the proc_macro crate is stable. The part that you can use from proc macros!

You can use it from outside proc-macros too.. kind of:

Rust code
// in `sample/src/main.rs`

extern crate proc_macro;

fn main() {
    println!(
        "some punct = {:#?}",
        proc_macro::Punct::new('+', proc_macro::Spacing::Joint)
    );
}
Shell session
$ cargo run --quiet
thread 'main' panicked at 'procedural macro API is used outside of a procedural macro', library/proc_macro/src/bridge/client.rs:346:17
note: run with `RUST_BACKTRACE=1` environment variable to display a backtrace

...it just panics at runtime. Which makes sense, because there's no proc macro server to talk to.

But everything else is behind unstable feature flags. If we try to use it:

Rust code
// in `sample/src/main.rs`

extern crate proc_macro;

struct MyServer {}

impl proc_macro::bridge::server::Types for MyServer {
    type FreeFunctions = ();
    type TokenStream = ();
    type TokenStreamBuilder = ();
    type TokenStreamIter = ();
    type Group = ();
    type Punct = ();
    type Ident = ();
    type Literal = ();
    type SourceFile = ();
    type MultiSpan = ();
    type Diagnostic = ();
    type Span = ();
}

fn main() {
    // muffin
}

A stable rustc will yell at us:

Shell session
$ cargo check
    Checking lhash v1.0.1
    Checking sample v0.1.0 (/home/amos/bearcove/sample)
error[E0658]: use of unstable library feature 'proc_macro_internals'
 --> src/main.rs:6:5
  |
6 |     type FreeFunctions = ();
  |     ^^^^^^^^^^^^^^^^^^^^^^^^
  |
  = note: see issue #27812 <https://github.com/rust-lang/rust/issues/27812> for more information

error[E0658]: use of unstable library feature 'proc_macro_internals'
 --> src/main.rs:7:5
  |
7 |     type TokenStream = ();
  |     ^^^^^^^^^^^^^^^^^^^^^^
  |
  = note: see issue #27812 <https://github.com/rust-lang/rust/issues/27812> for more information

(cut)

error[E0658]: use of unstable library feature 'proc_macro_internals'
 --> src/main.rs:5:6
  |
5 | impl proc_macro::bridge::server::Types for MyServer {
  |      ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  |
  = note: see issue #27812 <https://github.com/rust-lang/rust/issues/27812> for more information

For more information about this error, try `rustc --explain E0658`.
error: could not compile `sample` due to 13 previous errors

Of course there's escape hatches! We can simply add this to enable the unstable feature:

Rust code
// in `sample/src/main.rs`

#![feature(proc_macro_internals)]

// etc.

But... we can't use that on stable:

Rust code
$ cargo check
    Checking sample v0.1.0 (/home/amos/bearcove/sample)
error[E0554]: `#![feature]` may not be used on the stable release channel
 --> src/main.rs:1:12
  |
1 | #![feature(proc_macro_internals)]
  |            ^^^^^^^^^^^^^^^^^^^^

For more information about this error, try `rustc --explain E0554`.
error: could not compile `sample` due to previous error

Not unless we cheat a little and pretend we're rustc bootstrapping itself (and so, that we're allowed to use unstable features on stable):

Rust code
$ RUSTC_BOOTSTRAP=1 cargo check
    Checking sample v0.1.0 (/home/amos/bearcove/sample)
    Finished dev [unoptimized + debuginfo] target(s) in 0.07s

That alone is already an issue for the rust-analyzer project. It builds on rustc stable, and it would be nice if it stayed that way. Contributor experience is very important: it should be as easy as possible for someone to pick up the rust-analyzer codebase and add a feature, starting from this guide maybe.

But then there's another issue.

That interface is unstable for a reason. At the time of this writing, the stable version of rustc is 1.62:

Shell session
$ rustc --version
rustc 1.62.0 (a8314ef7d 2022-06-27)

And so that's what I used for the code samples. But if we switch to a recent nightly, everything falls apart:

Shell session
$ cargo +nightly-2022-07-23 check
   Compiling pm v0.1.0 (/home/amos/bearcove/pm)
    Checking lhash v1.0.1
    Checking sample v0.1.0 (/home/amos/bearcove/sample)
error[E0437]: type `TokenStreamBuilder` is not a member of trait `proc_macro::bridge::server::Types`
  --> src/main.rs:10:5
   |
10 |     type TokenStreamBuilder = ();
   |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ not a member of trait `proc_macro::bridge::server::Types`

error[E0437]: type `TokenStreamIter` is not a member of trait `proc_macro::bridge::server::Types`
  --> src/main.rs:11:5
   |
11 |     type TokenStreamIter = ();
   |     ^^^^^---------------^^^^^^
   |     |    |
   |     |    help: there is an associated type with a similar name: `TokenStream`
   |     not a member of trait `proc_macro::bridge::server::Types`

error[E0437]: type `Group` is not a member of trait `proc_macro::bridge::server::Types`
  --> src/main.rs:12:5
   |
12 |     type Group = ();
   |     ^^^^^^^^^^^^^^^^ not a member of trait `proc_macro::bridge::server::Types`

error[E0437]: type `Punct` is not a member of trait `proc_macro::bridge::server::Types`
  --> src/main.rs:13:5
   |
13 |     type Punct = ();
   |     ^^^^^^^^^^^^^^^^ not a member of trait `proc_macro::bridge::server::Types`

error[E0437]: type `Ident` is not a member of trait `proc_macro::bridge::server::Types`
  --> src/main.rs:14:5
   |
14 |     type Ident = ();
   |     ^^^^^^^^^^^^^^^^ not a member of trait `proc_macro::bridge::server::Types`

error[E0437]: type `Literal` is not a member of trait `proc_macro::bridge::server::Types`
  --> src/main.rs:15:5
   |
15 |     type Literal = ();
   |     ^^^^^^^^^^^^^^^^^^ not a member of trait `proc_macro::bridge::server::Types`

error[E0046]: not all trait items implemented, missing: `Symbol`
 --> src/main.rs:7:1
  |
7 | impl proc_macro::bridge::server::Types for MyServer {
  | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ missing `Symbol` in implementation
  |
  = help: implement the missing item: `type Symbol = Type;`

Some errors have detailed explanations: E0046, E0437.
For more information about an error, try `rustc --explain E0046`.
error: could not compile `sample` due to 7 previous errors

Because the proc_macro bridge changed!

Shock! Awe!

No! No shock. This is completely normal and expected β€” that's why it's unstable.

It's a rustc implementation detail, and there were some inefficiencies, which were recently tackled by Nika: here's one example.

Ah well, no big deal right? rust-analyzer just stops compiling and someone needs to fix it to get it to compile again?

No, very big deal in fact. rust-analyzer does not actually use the proc_macro sysroot crate. It has to build on stable, remember?

Ah! So what do they do?

Well! Because the proc macro bridge might change across rustc versions, and because rust-analyzer aims to be backwards-compatible for "a few rustc versions"... first it detects the rustc version a crate was compiled with, by reading the .rustc section of the dynamic library:

Shell session
$ llvm-objdump -s --section=.rustc target/debug/deps/libpm-e886d9f9eaf24619.so | head

target/debug/deps/libpm-e886d9f9eaf24619.so:    file format elf64-x86-64
Contents of section .rustc:
 0000 72757374 00000006 ff060000 734e6150  rust........sNaP
 0010 705900a2 02002049 3506bc09 30727573  pY.... I5...0rus
 0020 74000000 06000003 f023010d d0632031  t........#...c 1
 0030 2e36322e 30202861 38333134 65663764  .62.0 (a8314ef7d
 0040 20323032 322d3036 2d323729 c1000000   2022-06-27)....
 0050 01000609 646f5f74 68726963 65c10012  ....do_thrice...
 0060 00000140 5c00f501 00140000 0000200a  ...@\......... .

(rust-analyzer uses the object crate for this, so it works for ELF on Linux, Mach-O on macOS and PE on Windows).

You can see the version string in there: 1.62.0 (a8314ef7d 2022-06-27).

Then it parses that into this struct:

Rust code
// in `rust-analyzer/crates/proc-macro-api/src/version.rs`

#[derive(Debug)]
pub struct RustCInfo {
    pub version: (usize, usize, usize),
    pub channel: String,
    pub commit: Option<String>,
    pub date: Option<String>,
    // something like "rustc 1.58.1 (db9d1b20b 2022-01-20)"
    pub version_string: String,
}

And it uses that information to pick one of the ABI versions it supports:

Rust code
// in `rust-analyzer/crates/proc-macro-srv/src/abis/mod.rs`

        match (info.version.0, info.version.1) {
            (1, 58..=62) => {
                let inner = unsafe { Abi_1_58::from_lib(lib, symbol_name) }?;
                Ok(Abi::Abi1_58(inner))
            }
            (1, 63) => {
                let inner = unsafe { Abi_1_63::from_lib(lib, symbol_name) }?;
                Ok(Abi::Abi1_63(inner))
            }
            (1, 64..) => {
                let inner = unsafe { Abi_1_64::from_lib(lib, symbol_name) }?;
                Ok(Abi::Abi1_64(inner))
            }
            _ => Err(LoadProcMacroDylibError::UnsupportedABI),
        }

Wait a minute... one of the ABI it supports?

Well yeah! It needs to be backwards-compatible!

But how does it... how can it compile against different versions of the proc_macro sysroot crate?

Oh, it doesn't! And at first, it only supported one ABI. And since it needed to compile on stable, it didn't compile against the proc_macro sysroot crate at all. It simply used a manually copied-and-pasted version of it (with unstable feature gates removed).

Which, of course, broke every time the bridge changed, like in May 2021, for example.

So in July 2021, a scheme was devised β€” one I've just described: by reading the .rustc section, it could know which of its several hand-tuned copies of the proc_macro bridge code to use.

Oh.

And, you know. That kinda made the problem less visible for a while, for stable rustc versions, and even on nightly as long as the proc_macro bridge didn't change too much.

Before every stable release where the proc_macro bridge had changed, the following needed to be done:

Rust code
// in `rust-analyzer/crates/proc-macro-srv/src/abis/mod.rs`

//! # Adding a new ABI
//!
//! To add a new ABI you'll need to copy the source of the target proc_macro
//! crate from the source tree of the Rust compiler into this directory tree.
//! Then you'll need to modify it
//! - Remove any feature! or other things which won't compile on stable
//! - change any absolute imports to relative imports within the ABI tree
//!
//! Then you'll need to add a branch to the `Abi` enum and an implementation of
//! `Abi::expand`, `Abi::list_macros` and `Abi::from_lib` for the new ABI. See
//! `proc_macro_srv/src/abis/abi_1_47/mod.rs` for an example. Finally you'll
//! need to update the conditionals in `Abi::from_lib` to return your new ABI
//! for the relevant versions of the rust compiler

I've done it once, just to understand what it entailed, and depending on the proc_macro bridge changes, it can be a lot more work than this comment lets on.

For example, during the current cycle, Literal and Ident changed from being associated types (defined by the server) to being structs, re-using other associated types.

They now look like this:

Rust code
#[derive(Copy, Clone, Eq, PartialEq)]
pub struct Ident<Span, Symbol> {
    pub sym: Symbol,
    pub is_raw: bool,
    pub span: Span,
}

compound_traits!(struct Ident<Span, Symbol> { sym, is_raw, span });

#[derive(Clone, Eq, PartialEq)]
pub struct Literal<Span, Symbol> {
    pub kind: LitKind,
    pub symbol: Symbol,
    pub suffix: Option<Symbol>,
    pub span: Span,
}

Before, identifiers were interned. Now, symbols are interned, and they're referenced from both the new Ident struct and the new Literal struct.

This is a non-trivial change. Also, because Literal is now a struct, it comes with additional fields, like kind and suffix, that rust-analyzer's old Literal type simply did not have:

Rust code
// `rust-analyzer/crates/tt/src/lib.rs`

#[derive(Debug, Clone, PartialEq, Eq, Hash)]
pub struct Literal {
    pub text: SmolStr,
    pub id: TokenId,
}

And as part of the interface proc macro servers need to implement, you need to be able to convert between the bridge's type for literals, and the server's type for literal.

So, what do you do? Do you lex Literal::text to find the kind and suffix every time you need to perform the conversion? As it turns out, you can just stub them (with None and LitKind::err, respectively) and macro expansion still works well enough for rust-analyzer's use case.

But maybe in the future you won't! Maybe a larger API surface will be stabilized for the proc_macro crate, and you'll be able to access those fields from a proc macro, and then we won't be able to cheat anymore.

In the current 6-week cycle, changes like these were made several times. And every time, proc-macro support broke in rust-analyzer for rust nightly users.

Because nightly announces the same version for 6 weeks:

Rust code
$ rustc --version
rustc 1.64.0-nightly (848090dcd 2022-07-22)

...it's impossible (with rust-analyzer's multi ABI support scheme) to support multiple incompatible nightlies within the same cycle.

This made the situation completely untenable: even if you pinned your toolchain to a nightly version (using a toolchain file) that you knew worked with rust-analyzer at some point in time, any rust-analyzer upgrade could break it retroactively!

For example, at work, we used 2022-06-08. And for a while, it worked with rust-analyzer, thanks to some hand-tuned copy-pasting. But then more hand-tuned copy-pasting happened, and it broke for us!

And by break, here's what I mean: if you had a struct with a #[derive(serde::Serialize)] annotation on it? None of its code would be visible to rust-analyzer. It wouldn't see the relevant impl, so any usage of that type in conjunction with a (de)serializer would result in {unknown}.

No inlay annotations. No "Go to definition". No tab-completion, no nothing. Almost as if the code didn't exist at all.

Instead, you would see an error diagnostic along the lines of "proc macro server crashed". Or sometimes, failed to write request: Broken pipe (os error 32).

If you opened the "Rust Analyzer Server" output in vscode, you would see something like:

thread 'MacroExpander' panicked at 'assertion failed: `(left != right)`
  left: `0`,
 right: `0`', crates/proc-macro-srv/src/abis/abi_1_63/proc_macro/bridge/handle.rs:22:9
stack backtrace:
   0: rust_begin_unwind
             at /rustc/fe5b13d681f25ee6474be29d748c65adcd91f69e/library/std/src/panicking.rs:584:5
   1: core::panicking::panic_fmt
             at /rustc/fe5b13d681f25ee6474be29d748c65adcd91f69e/library/core/src/panicking.rs:143:14
   2: core::panicking::assert_failed_inner
   3: core::panicking::assert_failed
   4: proc_macro_srv::abis::abi_1_63::proc_macro::bridge::handle::InternedStore<T>::new
   5: proc_macro_srv::abis::abi_1_63::proc_macro::bridge::client::HandleStore<S>::new
   6: proc_macro_srv::abis::abi_1_63::proc_macro::bridge::server::run_server
   7: proc_macro_srv::abis::abi_1_63::proc_macro::bridge::server::<impl proc_macro_srv::abis::abi_1_63::proc_macro::bridge::client::Client<(proc_macro_srv::abis::abi_1_63::proc_macro::TokenStream,proc_macro_srv::abis::abi_1_63::proc_macro::TokenStream),proc_macro_srv::abis::abi_1_63::proc_macro::TokenStream>>::run
   8: proc_macro_srv::abis::abi_1_63::Abi::expand
   9: proc_macro_srv::abis::Abi::expand
  10: proc_macro_srv::ProcMacroSrv::expand
  11: proc_macro_srv::cli::run
note: Some details are omitted, run with `RUST_BACKTRACE=full` for a verbose backtrace.

Many, many such issues were opened against rust-analyzer's repository in the past couple months. Manual updates to the multi ABI scheme fixed things for some, and broke them for others.

It seemed there was no good way forward... until recently.

The long-term support plan

Since I had been suffering that issue along with many others, my curiosity was piqued when I read this June 2022 issue: RA should use the toolchain's proc_macro instead of copying library/proc_macro.

I had zero context back then, so I took a couple weeks to read up and chat with folks on various sides of the issue.

I knew rust-analyzer had joined the Rust organization back in Februrary 2022, and I knew it was distributed as a rustup component. You can see it in rustup components history: it used to be called rust-analyzer-preview, and now it's simply called rust-analyzer.

Which means you can do this:

Shell session
$ rustup component add rust-analyzer
info: component 'rust-analyzer' for target 'x86_64-unknown-linux-gnu' is up to date
$ rustup which rust-analyzer
/home/amos/.rustup/toolchains/nightly-2022-07-23-x86_64-unknown-linux-gnu/bin/rust-analyzer

Well. On nightly versions anyway. IIRC, it's slated for inclusion in the next stable release, but for now, it's not there:

Shell session
$ rustup +stable component add rust-analyzer
error: toolchain 'stable-x86_64-unknown-linux-gnu' does not contain component 'rust-analyzer' for target 'x86_64-unknown-linux-gnu'; did you mean 'rust-analysis'?

This seemed like a way out: surely the version of rust-analyzer shipped via rustup should be compatible with the rustc of the same channel, right?

Well, no! Because that rust-analyzer binary was simply built with that same multi-ABI scheme. What happened was:

And it would be built in Rust CI, so it could be shipped via rustup. But what would be built included several modified copies of the proc_macro bridge, with no guarantees whatsoever that they were compatible with the current rustc version.

To be sure that the rust-analyzer version is compatible with rustc, it has to use the sysroot proc_macro crate. That way, if the bridge internals change, it fails at compile time instead of crashing at runtime.

But... that additional "sysroot ABI", if you will, would be extremely difficult to maintain. It would have to be developed within the rust-analyzer repository, but could not be tested by rust-analyzer's CI, since it exclusively uses Rust stable.

So maintainers would need to test it locally, against specific rustc nightlies, then submit a PR to the rust-analyzer repository, wait for it to land, then submit a sync PR to the rust repository (bumping the rust-analyzer submodule version), and hope that the rust-analyzer component starts building again.

This could easily take a couple days, during which the rust-analyzer component would be missing from rustup channels.

But it would be better in some ways! At that point, you could look at the rustup components history page, and pick a nightly version that has a rust-analyzer component, and that version would work forever!

Except!

Remember how rust-analyzer has its own release schedule? Every monday like clockwork, there's a new set of features and bugfixes. And a couple hours later, VS Code installations all around the world download the latest version of the rust-analyzer VSCode extension from the VSCode marketplace, and... use the bundled binary.

The one that has the multi ABI scheme. And that doesn't have the "sysroot ABI". And that doesn't work on nightly.

Luckily, the VS Code rust-analyzer extension lets you override the path of the rust-analyzer binary to use, so if you just do:

JSON
{
    "rust-analyzer.server.path": "path/to/rustup/provided/rust-analyzer"
}

Then it would use the rustup-provided rust-analyzer binary, the one that would have the "sysroot ABI" and that is compatible with the rustc of the same channel.

But wait! Now you're missing out on all the cool new rust-analyzer features, because those are part of the rust-analyzer binary too! Whatever features it had back when that nightly was cut, you're stuck with. Which is not great.

But aha, what's this! There's another setting, one that allows setting just the "proc macro server" path:

JSON
{
    "rust-analyzer.procMacro.server": "path/to/rustup/provided/rust-analyzer"
}

And that ought to work! Except...

Oh boy.

...except that setting was kinda broken. For some reason, it interpreted its value as a path relative to the workspace root.

And there's two other issues with this. First: let's say there's a whole team working on the same cargo workspace, that depends on rust nightly, and most of the team is using VSCode: you'd want to commit that .vscode/settings.json file, right? So rust-analyzer works for everyone?

Well you can't! Because the actual path (if it worked) would be something like:

JSON
{
    "rust-analyzer.procMacro.server": "/home/amos/.rustup/toolchains/nightly-2022-07-23-x86_64-unknown-linux-gnu/bin/rust-analyzer"
}

And unless all your colleagues are named "amos" (cthulhu forbid), that just won't work.

Wait, doesn't rustup take care of that? You can do rustc --version and rustc +nightly --version for example? Why isn't there a rust-analyzer binary in $PATH?

Well, because not all tools distributed via rustup have this. "This", by the way, is a "proxy", and there's a PR to add one for rust-analyzer, but it hasn't landed yet.

Right, so we just wait for that to land, and then-

And then we still have problems, bear! The problems don't stop coming!

Because rust-analyzer assumed that its "proc macro server" could expand proc macros built with any "supported" rustc version, it only spawned one of them!

But you can have any number of cargo (or non-cargo!) workspaces open in VSCode (or other editors). And these workspaces might be using different rustc versions: they might each have a rust-toolchain.toml file with different toolchain.channel values.

If a single rust-analyzer proc-macro command was started, and it was started from some arbitrary directory, it would use the default toolchain's (probably stable) rust-analyzer binary, and it would only be compatible with that!

Instead, what we'd need is... to spawn one proc macro server per workspace. And so that's exactly what we did.

Ah, good! So that fixed everything, right?

Oh no bear. Far from it.

First off, someone still needed to actually add a "sysroot ABI". So I went ahead and did that.

But it was only a matter of days until that too, broke - and that code was not being tested in rust-lang CI, nor being built by rust CI, nor distributed via rustup. It was just one of many steps in my master plan to Never Have Proc Macros Break Ever Again.

For that "sysroot ABI" to be at all useful, it needed to be:

The first bullet point was by far the most work.

The idea is fairly simple: instead of rust-analyzer being a git submodule, it would become a git subtree.

Subtrees allow syncs in both directions: both repositories would have the full commit history for rust-analyzer, except RA's repo would have it under.. the root path, and Rust's repo would have it under src/tools/rust-analyzer.

This is exactly the same system used by clippy, for example.

Compare and contrast:

Syncs can happen in both directions, see:

rustfmt uses the same scheme: there's history in-tree and out-of-tree. Before that was the case, rustfmt was often missing from rustup. Now it's always present, and it always works!

So, after careful review, and several chats with current clippy and rust-analyzer maintainers, we went ahead and converted rust-analyzer to an in-tree tool (in other words, we moved from the "git submodule" model to the "git subtree" model).

To me, this was the moment rust-analyzer truly became president officially part of the Rust project. As of today, any commits to rust-lang/rust has to pass (a subset of) rust-analyzer tests, ensuring that it would never be broken again.

Rust CI builds the sysroot ABI, and rust-analyzer CI doesn't (it's a cargo feature that's simply not "default").

See, the problem here really wasn't technical: we had the technology to fix that all along. It was more of a communication problem.

rust-analyzer started relying on an unstable rustc interface because it had to. It had to do that in order to function as a code intelligence tool. And in fact, it's not the only one: IntelliJ Rust simply re-uses rust-analyzer's proc-macro-srv crate.

Because rust-analyzer needed to keep building on stable, wasn't distributed through rustup at the time (and even if it had been, the version used by most would've been the VS Code marketplace version β€” before that, a version downloaded by the VSCode extension from rust-analyzer's own GitHub repository), copying the proc_macro bridge sources seemed like a "good enough" compromise.

On the rust-lang/rust side of things, there were hard requirements too: the proc_macro bridge interface would not be stabilized. It would stay unstable to allow performance improvements and additional features (which have happened in the past couple months, thanks Nika!).

It's only after much discussion that we've reached that compromise: by turning rust-analyzer into a subtree, we would give the ability and the responsibility to rustc contributors to fix rust-analyzer's proc macro server, when breaking changes are made to the proc_macro bridge, with rust-analyzer developers' help if needed.

We've essentially worked out a "shared custody" agreement of a component nobody really wanted to create or maintain in the first place: rust-analyzer's proc-macro-srv crate.

But that brings us right back to our earlier issues with .vscode/settings.json and proc macro server paths and... being able to use all the latest rust-analyzer features, yet have a proc macros server that works with a given rust nightly.

At first, I came up with a plan that involved:

But it still wasn't ideal. rust-analyzer is not a default rustup component. So we would've had to teach the rust-analyzer binary (the one shipped with the VS Code extension, remember?) to:

It's not unheard of: rust-analyzer does that for the rust-src component, for example (see for yourself). It needs the standard library's sources to be able to index it and provide code intelligence for it.

Wouldn't it be better though... if you didn't have to install an extra component?

After all... we don't need the whole of rust-analyzer to be available: just the proc macro server part. Wouldn't it be neat if it was simply included alongside rustc? No matter whether it was installed through rustup, or your distribution's package manager, or your own custom build?

Well, we made that happen, too.

As I'm writing this, it was just merged into rust-lang/rust.

So, are we good?

All the pieces are in place β€” here's what's going to happen next.

In the next 24h, Rust nightly 2022-07-29 will be released (automatically).

It'll be the first version that includes a new executable in the sysroot:

Shell session
$ ls $(rustc --print sysroot)/libexec
cargo-credential-1password  rust-analyzer-proc-macro-srv

(Note: that one was handbuilt. I do not have time-travel powers.)

That executable behaves just like rust-analyzer's proc-macro subcommand.

Shell session
$ $(rustc --print sysroot)/libexec/rust-analyzer-proc-macro-srv
If you're rust-analyzer, you can use this tool by exporting RUST_ANALYZER_INTERNALS_DO_NOT_USE='this is unstable'.
If not, you probably shouldn't use this tool. But do what you want: I'm an error message, not a cop.

Well, okay, you just need to prove that you are, in fact rust-analyzer. Let's do a little role play:

Shell session
$ echo '{"ListMacros":{"dylib_path":"target/debug/deps/libpm-e886d9f9eaf24619.so"}}' | RUST_ANALYZER_INTERNALS_DO_NOT_USE='this is unstable' $(rustc --print sysroot)/libexec/rust-analyzer-proc-macro-srv
{"ListMacros":{"Ok":[["do_thrice","FuncLike"]]}}

That's one piece.

The other has already landed in rust-analyzer. (With a discreet follow-up nobody needs to know about).

Because rust-analyzer now spins up one proc macro server per workspace, for every workspace, it'll simply look into the sysroot and see if it finds a rust-analyzer-proc-macro-srv binary:

Rust code
// in `rust-analyzer/crates/rust-analyzer/src/reload.rs`

                        if let ProjectWorkspace::Cargo { sysroot, .. } = ws {
                            tracing::info!("Found a cargo workspace...");
                            if let Some(sysroot) = sysroot.as_ref() {
                                tracing::info!("Found a cargo workspace with a sysroot...");
                                let server_path = sysroot
                                    .root()
                                    .join("libexec")
                                    .join("rust-analyzer-proc-macro-srv");
                                if std::fs::metadata(&server_path).is_ok() {
                                    tracing::info!(
                                        "And the server exists at {}",
                                        server_path.display()
                                    );
                                    path = server_path;
                                    args = vec![];
                                } else {
                                    tracing::info!(
                                        "And the server does not exist at {}",
                                        server_path.display()
                                    );
                                }
                            }
                        }

So, if there's no last-minute catastrophe, on Monday August 1st, a new stable rust-analyzer version will drop, and it will be compatible with all rust nightlies starting from 2022-07-29.

(I won't have a computer on me at this date, so, fingers crossed!)

And to think all it took was spending two weeks talking to 10+ folks over GitHub, Zulip and Discord, and landing these PRs!

What's next?

I'm extremely pleased with the solution we all landed on. It'll work out-of-the-box for all nightlies moving forward, rust-analyzer won't break retroactively, it'll work regardless of how rustc is installed, unless distro package maintainers explicitly delete the binary from their package (please don't do that, I will cry).

The proc_macro bridge changes can continue without fear of breaking rust-analyzer (and there's some changes planned fairly soon!).

Over time, the JSON interface between rust-analyzer and rust-analyzer-proc-macro-srv might need to evolve - since it's essentially a network protocol (only, over stdio), any of the usual methods can be used to achieve that - as long as it's backwards-compatible for a bit!

I still think some of the shortcuts taken in rust-analyzer's proc macro server implementation are.. just wrong. I think it should track spans. I think it should validate and nfc-normalize identifiers, like rustc does. And I think it should lex literals and keep track of their types/suffixes.

Right now the only way you can see that rust-analyzer cheats is by using the Debug impl of these types to generate the output of your proc-macro, which hopefully nobody is doing. But later on, this might not be the case! Because rust-analyzer is now a subtree, all options are open.

I would love to see all rust-analyzer tests pass on all supported Rust platforms. Even after fixing tests so they don't write to the source directory (which is read-only in Rust CI, as it should), and fixing the code so it passes the stricter lints, we noticed that some tests failed due to, hang on to your socks, different iteration order for FxHashSet and FxHashMap on i686.

(I tried fixing some but there were more and we decided to just disable those tests in rust CI for now.)

Finally, although I remain convinced that moving to git subtree was the right thing to do, and that all official Rust projects really should be doing that (looking at you Miri), the first-time setup is, uh, a bit of a hassle.

It's all detailed in the Clippy book. You need to:

What's that mapping for? git-subtree operates on two repositories at once: the "top" repository (rust) and the "subtree" repository (rust-analyzer, rust-clippy, rustfmt, etc.).

Because the whole history of those projects is contained in both, the first time you run git subtree pull or git subtree push, it needs to find out pairs of matching commits from both repositories. After that, it knows which commits are missing and it's able to perform syncs in either direction.

This was a pretty bad surprise, but not a blocker, considering that it's a one-time thing, and that in practice, very few people actually perform syncs in either direction (tragedy of the commons and all that).

Still, since the issue was raised, I started doing a bit of research into faster alternatives. After all, if git-filter-repo is able to rewrite 25K commits in a couple seconds so the repo is rooted at a different path (./src/tools/rust-analyzer/ instead of ./), couldn't we work something out?

As for rust-analyzer, well, my hope is that in a few months, all rustc versions that rust-analyzer considers "supported" will have that rust-analyzer-proc-macro-srv binary, and we'll be able to retire the "multi ABI compatibility scheme" for good, deleting a bunch of code that really should have never existed in the first place.

Thanks

I would like to thank, in random order: Eduard-Mihai Burtescu, Lukas Wirth, Joshua Nelson, Edwin Cheng, LaurenΘ›iu Nicola, Jonas Schievink, bjorn3, Pietro Albini, Florian Diebold, Jade Lovelace, Aleksey Kladov, Jubilee, Oli Scherer and Mark Rousskov for their contribution to this endeavor.

Bye for now

Coordinating and landing all these changes has been a lot of fun (meeting new people! Learning new things! Making an impact!), but also exhausting.

I'm going to take a small break from trying to land stuff in two different projects at once and enjoy a working rust-analyzer, celebrating every monday on the dot as new features roll in.

In the meantime, take care, and enjoy!