How is this different from backtracking? You're doing a depth-first search over possible interpretations. The grammar is just expressed in the type system instead of usual spec formats.
Critiques in other comments are accurate. This is a tooling nightmare, but also probably a nightmare to read. Consider an expression like
2026 March 10 to 13
What's the binding precedence? Does this mean March 10 through March 13, or midnight to 1 PM on March 10th? I think this breaks down outside of trivial examples that are better achieved in other ways.Yes, technically this is a form of backtracking, similar to what a parser does. The key difference is that the search is drastically constrained by the type system: reductions are only attempted where the types actually support a binding operator. Unlike a parser exploring all grammar possibilities, this mechanism prunes most candidates automatically, so the compiler efficiently "solves" the expression rather than blindly exploring every syntactic alternative.
Here is the high-level explanation of the mechanism:
https://github.com/manifold-systems/manifold/tree/master/man...
But the short answer is that it’s not parser-style backtracking over a grammar.
The Java parser still produces a normal AST for the sequence of tokens. What happens afterward is a type-directed binding phase where adjacent expressions may bind if their types agree on a binding operator. The compiler effectively reduces the expression by forming larger typed expressions until it reaches a stable form.
The algorithm favors left associativity, but since a type can implement the binding operator as either the left or right operand, the overall structure of the expression can emerge in different ways depending on the participating types.
So rather than exploring grammar productions, the compiler is solving a set of type-compatible reductions across the expression.
For example:
2026 March 10
reduces roughly like this: (2026 (March 10))
→ March.postfixBind(2026) // → LocalYearMonth
→ [retreat] // → no binding with 10
→ March.prefixBind(10) // → LocalMonthDay
→ .postfixBind(2026) // → LocalDate
And if `Month` binds with `Range<Integer>`: 2026 March 10 to 13
can reduce as: (2026 (March ((10 to) 13)))
The meaning is therefore determined entirely by which types participate in binding e.g., `LocalDate`, `Month`, `Integer`, `Range`, etc. and which reductions they define.If a competing interpretation exists but the types don’t support the necessary bindings, it simply never forms.
In that sense it behaves less like a traditional parser and more like a typed reduction system layered on top of the Java AST.
2025 July 19 // → LocalDate
300M m/s // → Velocity
1 to 10 // → Range<Int>
Schedule Alice Tues 3pm // → CalendarEvent
That's the idea behind binding expressions — a compiler plugin I built to explore what it would mean if adjacency had operator semantics. It lets adjacent expressions bind based on their static types, forming new expressions through type-directed resolution.Details here: https://github.com/manifold-systems/manifold/blob/master/doc...
The formal name for the “empty” binary infix operator that gets implied in the AST when doing this, is the “juxtaposition” (or “juxtapose”, or “juxt”) operator. The implicit multiplication operator between `3` and `a` in the polynomial expression `3a + 4`, and the implicit function-application operator in the Lambda-calculus expression `f x y`, are both instances of an implied juxtaposition operator (with different semantics for it in each of the two cases, as befits each type of algebra/calculus.)
This seems like a great attempt. I would be worried about how much parsing and backtracking might be required to infer the infix precedence in a totally general system (like garden-path sentences[1]) or actually ambiguous parse trees (which is cured by adopting some rule like right precedence and parens, but what rule you pick makes some 'natural language' constructions work over others).
2 `plus` 3
rather than plus 2 3 -- infix
select a <!!!> b;
-- prefix
select <||> a;
A lot of custom types end up using this [1]. select @-@ '[(0,0),(1,0),(1,1)]'::path;
-- 2
[0] https://www.postgresql.org/docs/current/sql-createoperator.h...
[1] https://www.postgresql.org/docs/current/functions-geometry.h...You could of course affix all lemmata with structural information, as free word order languages do, but that's introducing syntactic structure via the backdoor.
[1] https://www.researchgate.net/publication/2743686_Are_Ours_Re...
So, yes, it can be done and it was done. Yes, expressiveness rises. No, reading comprehension of such languages does not suffer. Yes, it has to have a lot of scaffolding.
If you added a function to the examples, you could do a few of them, e.g.:
2025 July 19 date
299.8 M m / s velocity
But even this breaks down when you get to something like “Meet Alice Tuesday at 3pm”. Sure, you could contort things to make it resemble the concept, but it’d be a stretch at best.
It’s similar for the human reader: The examples are only intelligible to the reader incidentally, due to the names used and some natural-text conventions. In the general case, you have a seemingly random token sequence where you have no idea what binds to what, without looking up the type definitions or having an IDE present the expression in some structured way again.
Furthermore, in typical code you don’t have the case of constant values so often. You’ll rather have things like:
Well, yes and no. During AST building a binding expression resolves as an untyped polyadic expression. Only later during the compiler's type attribution phase does a binding expression's structure resolve based on the operand types.
https://github.com/manifold-systems/manifold/tree/master/man...
> in typical code you don’t have the case of constant values so often.
Agreed. It's not really useful with inlined expressions:
but if you write it like: But, honestly, I can't say I personally use it that way ;)Initially, I wrote it as a science extension to Java: https://github.com/manifold-systems/manifold/tree/master/man...