Show HN: Glowstick – type level tensor shapes in stable rust
github.com47 points by bietroi a day ago
47 points by bietroi a day ago
Hi HN,
In the past few years I've become more interested in machine learning. Since I'm sure the same is true for many here, I wanted to share this project I've been working on: glowstick uses type-directed metaprogramming to keep track of tensor shapes in Rust's type system and determine which operations are permitted or not at compile time.
I find Rust has a lot of strengths when it comes to ML applications, but waiting until runtime to find shape related issues feels a bit strange since normally I don't run the code all that often while developing. Given Rust has fancy types available, I figured I'd try my hand at using them to address this.
I've added integration crates for the two ML frameworks I use most frequently, candle and burn, and included examples of implementing llama 3.2 in each using typed shapes for much of the model internals and inference loop. Mixtures of static and dynamic dimensions should be supported well enough for most applications at this point, though there are of course still improvements to be made.
Any feedback is appreciated!
Could you take a look at Barry Jay's shape theory. https://web.archive.org/web/20111015133833/http://www-staff.... This was used in his shape aware language FiSh, for dealing with multidimensional arrays. Shape compatibilities were statically type checked, if I recall correctly. Shapes were also used to optimize the loops. [Programming in FISh]
https://link.springer.com/article/10.1007/s100090050037 [Towards Dynamic Shaping]
https://www.researchgate.net/publication/265975794_Towards_D... It's cool to see that others have explored some of these concepts more generally, thanks for the links! The glowstick shape is also a type-level list of integers, and I could definitely see how other shapes might be useful in different situations. Doing this sort of thing in Rust is a bit of a stretch for sure, but it makes the outcome easy to apply which is nice. I wonder… I know Eigen has some tricks it can do when the size of a matrix is known beforehand. The obvious example, 4x4 matrix inverse gets special treatment. I assume they also be smart about loop unrolling, that sort of stuff. Anything similar in here? If not—actually, optimizing compilers are pretty okay nowadays anyway. I wonder if you’ve tried just seeing what Rust will do automatically with different optimization levels? Glowstick just provides the shape types and associated traits as a layer you can put on top of another tensor implementation. Since it's just verifying shapes and forwarding the operations to the underlying tensor (e.g. from candle/burn), I don't think there's any great way to get performance benefits from these integrations. It's mainly about the developer experience- getting errors at compile time vs runtime, checking shapes, etc. That being said, it seems reasonable that you could make some optimizations like this if you had deeper integration of these types with a framework or similar. It's not something I've explored personally, sounds interesting though.
srean - a day ago
bietroi - 13 hours ago
bee_rider - a day ago
bietroi - 15 hours ago