What are some cool and obscure data structure you know of?
from protein@programming.dev to programming@programming.dev on 01 Jan 2026 07:08
https://programming.dev/post/43298192

#programming

threaded - newest

felsiq@piefed.zip on 01 Jan 2026 07:13 next collapse

Conflict free replicated data types, I don’t know if I’d call them obscure but they’re definitely cool and less often used. They’re for shared state across computers, like in collaborative apps

tyler@programming.dev on 01 Jan 2026 08:14 next collapse

They were pretty obscure until recently! I would say most people still don’t know about them.

Pissmidget@lemmy.world on 01 Jan 2026 15:31 next collapse

From just the name my mind instantly thought of the conflict as “conflict diamonds”, and I began to wonder what constitutes a conflict free boolean or integer.

If anyone wants to take a crack at writing up why primitives are unfortunate, and we should move on to new "conflict free data types"™ I will cheer you on!

Also, very interesting read about actual conflict free replicated days types. Cheers!

BackgrndNoize@lemmy.world on 01 Jan 2026 16:02 collapse

This sounds like document collaboration software like Google sheets where multiple people can edit a document at the same time

Transform2942@lemmy.ml on 01 Jan 2026 16:22 collapse

Yes exactly, collaborative editing is probably the number one use case people actually use them for. I doubt Google docs actually uses a real CRDT behind the scenes but they do have some big brains over there in the chocolate factory

marlinz@sueden.social on 01 Jan 2026 07:20 next collapse

@protein

Finger Tree!

A persistent, purely functional workhorse. Amortized O(1) access at both ends, O(log n) concatenation/splitting.

It generalizes elegantly to build sequences, priority queues, and more. Powers Haskell's main Data.Sequence. A functional programmer's secret weapon.

Vorpal@programming.dev on 01 Jan 2026 14:36 collapse

On paper they are efficient. In practise, all pointer based data structures (linked lists, binary trees, etc) are slow on modern hardware. And this effect is more important than the complexity in practise for most practical high performance code.

You are far better off with linear access where possible (e.g. vectors, open addressing hash maps) or if you must have a tree, make the fan-out factor as large as possible (e.g. btrees rather than binary trees).

Now, I don’t know if Haskell etc affords you such control, I mainly code in Rust (and C++ in the past).

Also see this old thread from 2016 on hacker news about this very topic: news.ycombinator.com/item?id=13263275

coherent_domain@infosec.pub on 01 Jan 2026 15:59 next collapse

I don’t know if Haskell etc affords you such control

You can have immutable arrary with vectors, but to mutate them you will need to wrap your action in a Monad. It even supports unboxed values.

hackage.haskell.org/package/vector

But I agree boxed default actually causes a lot of performance overhead in many high-level languages.

marlinz@sueden.social on 01 Jan 2026 16:30 collapse

@Vorpal

Totally fair point, thanks for calling that out.

When I mentioned finger trees I was thinking more about the *functional* side (persistence, elegant composition, Haskell/Data.Sequence style usage) than raw performance on real hardware.

In performance‑critical code your argument for cache‑friendly linear structures and wide trees absolutely makes sense, and I appreciate the reminder to think about actual access patterns and hardware effects, not just asymptotic complexity.

Vorpal@programming.dev on 01 Jan 2026 19:23 next collapse

I think a lot of modern software is bloated. I remember when GUI programs used to fit on a floppy or two. Nowdays we have bloated electron programs taking hundreds of MB of RAM just to show a simple text editor, because it drags a whole browser with it.

I love snappy software, and while I don’t think we need to go back to programs fitting on a single floppy and using hundreds of KB of RAM, the pendulum does need to swing back a fair bit. I rewrote some CLI programs in the last few years that I found slow (one my own previously written in Python, the other written in C++ but not properly designed for speed). I used Rust, which sure helped compared to Python, but the real key was thinking carefully about the data structures used up front and designing for performance. And lots of profiling and benchmarking as I went along.

The results? The python program was sped up by 50x, the C++ program by 320x. In both cases it changed these from “irritating delay” to “functionally instant for human perception”.

The two programs:

And I also rewrote a program I used to manage Arch Linux configs (written in bash) in Rust. I also added features I wanted so it was never directly comparable (and I don’t have numbers), but it made “apply configs to system” take seconds instead of minutes, with several additional features as well. (github.com/VorpalBlade/paketkoll/…/konfigkoll)

Oh and want a faster way to check file integrity vs the package manager on your Linux distro? Did that too.

Now what was the point I was making again? Maybe I’m just sensitive to slow software. I disable all animations in GUIs after all, all those milliseconds of waiting adds up over the years. Computers are amazingly fast these days, we shouldn’t make them slower than they have to be. So I think far more software should count as performance critical. Anything a human has to wait for should be.

Faster software is more efficient as well, using less electricity, making your phone/laptop battery last longer (since the CPU can go back to sleep sooner). And saves you money in the cloud. Imagine if you could save 30-50% on your cloud bill by renting fewer resources? Over the last few years I have seen multiple reports of this happening when companies rewrite in Rust (C++ would also do this, but why would you want to move to C++ these days?). And hyperscalers save millions in electricity by optimising their logging library by just a few percent.

Most modern software on modern CPUs is bottlenecked on memory bandwidth, so it makes sense to spend effort on data representation. Sure start with some basic profiling to find obvious stupid things (all non-trivial software that hasn’t been optimised has stupid things), but once you exhausted that, you need to look at memory layout.

(My dayjob involves hard realtime embedded software. No, I swear that is unrelated to this.)

AbelianGrape@beehaw.org on 02 Jan 2026 05:38 collapse

I don’t mean this in a bad way; did an AI write this comment?

duckythescientist@sh.itjust.works on 01 Jan 2026 07:30 next collapse

I’m also not sure if this is obscure, but Bloom Filters! It’s a structure that you can add elements to then ask it if it has seen the element before with the answer being either “no” or “probably yes”. There’s a trade-off between confidence of a “probably yes”, how many elements you expect to add, and how big the Bloom Filter is, but it’s very space and time efficient. And it uses hash functions which always make for a fun time.

FizzyOrange@programming.dev on 01 Jan 2026 09:46 next collapse

Obscure 10 years ago maybe. These days there have been so many articles about them I bet they’re more widely known than more useful and standard things like prefix trees (aka tries).

sukhmel@programming.dev on 01 Jan 2026 11:45 collapse

Relevant xkcd

in Randall's words

> Sometimes, you can tell Bloom filters are the wrong tool for the job, but when they’re the right one you can never be sure.

Nomad@infosec.pub on 01 Jan 2026 08:02 next collapse

Merkle trees

tyler@programming.dev on 01 Jan 2026 08:14 collapse

Aren’t these one of the first structures you learn about in any comp sci course? Still good to know but not sure it’s obscure.

coherent_domain@infosec.pub on 01 Jan 2026 16:14 next collapse

Not disagreeing with you, but I find it funny that this is the only data structure I have not heard of in this entire thread 🤣

tyler@programming.dev on 02 Jan 2026 17:07 collapse

Haha that is pretty funny. 😂

Nomad@infosec.pub on 01 Jan 2026 18:57 collapse

Old tech is more like it. Good basics but you wouldn’t code in ASM must of the time even if you learned it.

TeamAssimilation@infosec.pub on 01 Jan 2026 08:08 next collapse

Maybe not that obscure, but Joe Celko’s Nested Set Model gave me exactly what I needed when I learned of it: fast queries on seldom-changing hierarchical database records.

Updates are heavy, but the reads are incredibly light.

en.wikipedia.org/wiki/Nested_set_model

notabot@piefed.social on 01 Jan 2026 12:13 collapse

I came here to mention these too. One addition that can be helpful in large trees is to add a depth attribute to each node so that you can easily limit the depth of subtree you retrieve.

litchralee@sh.itjust.works on 01 Jan 2026 08:16 next collapse

IMO, circular buffers with two advancing pointers are an awesome data structure for high performance compute. They’re used in virtualized network hardware (see virtio) and minimizing Linux syscalls (see io_uring). Each ring implements a single producer, single consumer queue, so two rings are usually used for bidirectional data transfer.

It’s kinda obscure because the need for asynchronous-transfer queues doesn’t show up that often unless dealing with hardware or crossing outside of a single CPU. But it’s becoming relevant due to coprocessors (ie small ARM CPUs attached to a main CPU) that process offloaded requests and then quickly return the result when ready.

xthexder@l.sw0.com on 01 Jan 2026 17:48 collapse

One cool trick that can be used with circular buffers is to use memory mapping to map the same block of memory to 2 consecutive virtual address blocks. That way you can read the entire contents of the buffer as if it was just a regular linear buffer with an offset.

tiredofsametab@fedia.io on 01 Jan 2026 08:13 next collapse

Not necessarily obscure, but I don't think Tries get enough love.

Edit: I can't spell

muzzle@lemmy.zip on 01 Jan 2026 09:45 collapse

Tris?

catchy_name@feddit.it on 01 Jan 2026 10:21 collapse

They likely meant tries / prefix trees.

tiredofsametab@fedia.io on 01 Jan 2026 10:27 next collapse

Yep. I can't spell today

BonkTheAnnoyed@lemmy.blahaj.zone on 01 Jan 2026 14:52 collapse

That’s really neat!

muzzle@lemmy.zip on 01 Jan 2026 11:15 collapse

Trie!

MonkderVierte@lemmy.zip on 01 Jan 2026 10:30 next collapse

Archivemount

xep@discuss.online on 01 Jan 2026 12:39 next collapse

How about this variation of linked lists? data-structures-in-practice.com/intrusive-linked-…

FizzyOrange@programming.dev on 01 Jan 2026 23:22 collapse

Quite well known now because Rust-haters like to point out how they’re awkward to use in Rust.

palordrolap@fedia.io on 01 Jan 2026 14:01 next collapse

An ultimately doomed one that existed in Perl for a while was the pseudohash. They were regular integer-indexed arrays that could be accessed as though they were hashes (aka associative arrays / dictionaries). They even made it into the main Perl books at the time as this awesome time saving device. Except they weren't.

I did a quick web search just now and someone did a talk about why they weren't a great idea and they tell it better than I could; Link: https://perl.plover.com/classes/pseudohashes/

The supplied video doesn't have great sound quality, and it might be better to just click through the slides under Outline at the bottom there.

Vorpal@programming.dev on 01 Jan 2026 14:17 next collapse

XOR lists are obscure and cursed but cool. And not useful on modern hardware as the CPU can’t predict access patterns. They date from a time when every byte of memory counted and CPUs didn’t have pipelines.

(In general, all linked lists or trees are terrible for performance on modern CPUs. Prefer vectors or btrees with large fanout factors. There are some niche use cases still for linked lists in for example kernels, but unless you know exactly what you are doing you shouldn’t use linked data structures.)

EDIT: Fixed spelling

xthexder@l.sw0.com on 01 Jan 2026 16:15 next collapse

I came up with a kind of clever data type for storing short strings in a fixed size struct so they can be stored on the stack or inline without any allocations.
It’s always null-terminated so it can be passed directly as a C-style string, but it also stores the string length without using any additional data (Getting the length would normally have to iterate to find the end).
The trick is to store the number of unused bytes in the last character of the buffer. When the string is full, there are 0 unused bytes and the size byte overlaps the null terminator.
(Only works for strings < 256 chars excluding null byte)

Implementation in C++ here: github.com/frustra/…/InlineString.hh

Edit: Since a couple people don’t seem to understand the performance impact of this vs regular std::string, here’s a demo: godbolt.org/z/34j7obnbs This generates 10000 strings like “Hello, World! 00001” via concatenation. The effect is huge in debug mode, but still has performance benefits with optimizations turned on:

With -O3 optimization
std::string: 0.949216ms
char[256] with strlen: 0.88104ms
char[256] without strlen: 0.684734ms

With no optimization:
std::string: 3.5501ms
char[256] with strlen: 0.885888ms
char[256] without strlen: 0.687733ms

(You may need to run it a few times to get sample numbers due to random server load on godbolt)
Changing the buffer size to 32 bytes has a negligible performance improvement over 256 bytes in this case, but might be slightly faster due to the whole string fitting in a cache line.
FizzyOrange@programming.dev on 01 Jan 2026 23:20 next collapse

Interesting idea, but your trick is never really going to help (you can store up to 255 bytes instead of 254). Also always using 256 bytes for every string seems wasteful.

I think LLVM’s small string optimisation is always going to be a better option: joellaity.com/2020/01/31/string.html

xthexder@l.sw0.com on 01 Jan 2026 23:41 collapse

22 characters is significantly less useful than 255 characters. I use this for resource name keys, asset file paths, and a few other scenarios. The max size is configurable, so I know that nothing I am going to store is ever going to require heap allocations (really bad to be doing every frame in a game engine).

I developed this specifically after benchmarking a simpler version and noticed a significant amount of time being spent in strlen(), and it had real benefits in my case.
Admittedly just storing a struct with a static buffer and separate size would have worked pretty much the same and eliminated the 255 char limitation, but it was fun to build.

FizzyOrange@programming.dev on 02 Jan 2026 22:00 collapse

22 characters is significantly less useful than 255 characters.

You can still use more than 22 characters; it just switches to the heap.

nothing I am going to store is ever going to require heap allocations

I would put good money that using 256 bytes everywhere is going to be slower overall than just using the heap when you need more than 22 characters. 22 is quite a lot, especially for keys. ThisReallyLongKey is still only 17.

xthexder@l.sw0.com on 03 Jan 2026 05:33 collapse

I don’t use 256 bytes everywhere. I use a mix of 64, 128, and 256 byte strings depending on the specific use case.
In a hot path, having the data inline is much more important than saving a few hundred bytes. Cache efficiency plus eliminating heap allocations has huge performance benefits in a game engine that’s running frames as fast as possible.

FizzyOrange@programming.dev on 03 Jan 2026 22:22 collapse

having the data inline

It’s not as simple as that, depending on the architecture. Typically they would fetch 64-byte cache lines so your 128 bytes aren’t going to be magically more cached than 128 bytes on the heap.

Avoiding allocations may help but also maybe not. This is definitely in “I don’t believe it until I see benchmarks” realm. I would be really really surprised if the allocation cost was remotely bad enough to justify the “sorry your file is too long” errors.

xthexder@l.sw0.com on 03 Jan 2026 23:04 collapse

Check out the benchmark I edited in to my original post. These are not user-provided strings in my case.

anton@lemmy.blahaj.zone on 03 Jan 2026 09:45 collapse

I came up with a kind of clever data type for storing short strings in a fixed size struct so they can be stored on the stack or inline without any allocations.

C++ already does that for short strings while seamlessly switching to allocation for long strings.

It’s always null-terminated so it can be passed directly as a C-style string, but it also stores the string length without using any additional data (Getting the length would normally have to iterate to find the end).

Also the case in the standard library

The trick is to store the number of unused bytes in the last character of the buffer. When the string is full, there are 0 unused bytes and the size byte overlaps the null terminator.

Iirc, that trick was used in one implementation but discontinued because it was against the standard.

(Only works for strings < 256 chars excluding null byte)

If you need a niche for allocated string you can get to 254 but the typical choice seem to be around 16.

xthexder@l.sw0.com on 03 Jan 2026 16:32 collapse

C++ already does that for short strings

I’ve already been discussing this. Maybe read the rest of the thread.

Also the case in the standard library

I think you’re missing the point of why. I built this to be a nearly drop in replacement for the standard string. If this wasn’t the case it would need to do even more processing and work to pass the strings to anything.

discontinued because it was against the standard.

Standards don’t matter for an internal type that’s not exposed to public APIs. I’m not trying to be exactly compatible with everything under the sun. There’s no undefined behavior here so it’s fine

solomonschuler@lemmy.zip on 01 Jan 2026 17:24 next collapse

skiplists are interesting data structures. The underlying mechanism is it’s a 2-dimensional probabilistic linked list with some associated height ‘h’ that enables skipping of nodes through key-value pairs. So, compared to a traditional linked list that uses a traversal method to search through all values stored. A skip list starts from the maxLevel/maxheight, determines if “next” points to a key greater than the key provided or a nullptr, and moves down to the level below it if it is. This reduces the time complexity from O(1) with a linked list to O(N) where N Is the maxLevel.

The reason behind why its probabilistic (in this case using a pseudo random number) is because its easier to insert and remove elements, otherwise (if you went with the idealized theoretical form) you would have to reconstruct the entire data structure each and every time you want to add/remove elements.

In my testing when adding 1,000,000 elements to a skiplist it reduced from 6s search with a linked list to less than 1s!

Gobbel2000@programming.dev on 01 Jan 2026 17:26 next collapse

The CSR (compressed sparse row) format is a very simple but efficient way of storing sparse matrices, meaning matrices with a large amount of zero entries, which should not all occupy memory. It has three arrays: one holds all non-zero entries in order, read row by row, the next array contains the column indices of each non-zero element (and therefore has the same length as the first array), the third array indices into the first array for the first element of each row, so we can tell where a new row starts.

On sparse matrices it has optimal memory efficiency and fast lookups, the main downside is that adding or removing elements from the matrix requires shifting all three arrays, so it is mostly useful for immutable data.

jxk@sh.itjust.works on 03 Jan 2026 18:12 collapse

Oh yeah that’s a good one

And also, if you’re representing a 0/1 matrix, you can just do away with the first column altogether.

Gobbel2000@programming.dev on 03 Jan 2026 22:55 collapse

Right, which occurs in particular when dealing with graphs, which are basically matrices and usually sparse. Large graphs are what I used this format for, however I also needed edge weights, so the first column was still required for that.

myfavouritename@beehaw.org on 01 Jan 2026 17:45 next collapse

I get way more use out of Doubly Connected Edge Lists (DCEL) than I ever thought I would when I first learned about them in school.

When I want to render simple stuff to the screen, built-in functions like ‘circle’ or ‘line’ work. But for any shapes more complicated than that, I often find that it’s useful to work with the data in DCEL form.

idunnololz@lemmy.world on 02 Jan 2026 01:49 next collapse

I personally don’t think it’s that obscure but I have never seen this used in production code that I didn’t write: the linked hash map or ordered hash map.

jesale@mastodon.social on 02 Jan 2026 06:05 next collapse

@idunnololz @protein what's the goal of an ordered hashmap?

mcmodknower@programming.dev on 02 Jan 2026 16:31 collapse

Having a fixed iteration order, but an amortized O(1) access time.

For example if you have a map of uuids to players in a game, and want to use the same map for doing operations to all of them in the order they joined.

3abas@lemmy.world on 02 Jan 2026 06:12 collapse

I’ve used it twice that I can recall.

dustletter@piefed.blahaj.zone on 02 Jan 2026 06:25 next collapse

Skew binary trees. They’re an immutable data structure combining the performance characteristics of lists (O(1) non-amortized push/pop) and b-trees (log(N) lookup and updates)
They use a sequence of complete trees, cleverly arranged using skew binary numbers so that adding an element never causes cascading updates.
In practice they’re superseded by relaxed radix balanced trees.

Atlas_@lemmy.world on 02 Jan 2026 08:11 next collapse

Fibonacci heaps are pretty cool. Not used very often b/c they’re awful to implement, but better complexity than many other heaps.

Also Binary Lifting is closer to an algorithm than a data structure but it’s used in Competitive Programming a fair bit, and isn’t often taught: cp-algorithms.com/graph/lca_binary_lifting.html

And again closer to an algo tham a data structure, but Sum over Subsets DP in 3^n also has a cool little bit of math in it: cp-algorithms.com/algebra/all-submasks.html

QueenMidna@lemmy.ca on 02 Jan 2026 17:35 next collapse

I’ve been knee-deep in these lately so I’m a big fan

Theta sketches!

Do you want to approximately count a large volume of items, but save the state for later so you can UNION , INTERSECT and even DIFF them? Then Thetas are right for you!

Or basically anything in the Apache Datasketches lbrary.

runeko@programming.dev on 02 Jan 2026 17:35 next collapse

Dewey decimal

JackbyDev@programming.dev on 02 Jan 2026 22:29 next collapse

B trees are cool but not obscure necessarily. I didn’t learn about them in college. It sounds like binary tree and it’s similar but it’s different. It’s a data structure to take advantage of the way disk reads work.

Trigger2_2000@sh.itjust.works on 03 Jan 2026 00:06 collapse

AVL Trees are neat AVL Tree

Trigger2_2000@sh.itjust.works on 03 Jan 2026 00:19 next collapse

Not really a data structure per say, but just knowing LISP and the interesting structures it uses internally.

The results of LISP operations CAR, CDR, CADR and the other one I can’t remember now.

InformationEntropy@programming.dev on 03 Jan 2026 10:48 next collapse

Locality Sensitive Hashing. Not sure if it’s more algorithm than data structure, or even if it’s obscure, but it is very cool.

Eufalconimorph@discuss.tchncs.de on 04 Jan 2026 01:36 collapse

Not exactly a datastructure alone, but bitslicing is a neat trick to turn some variable-time operations into constant-time operations. Used in cryptography for “substitution box” (S-box) operations, which can otherwise leak secrets via data-dependent timing variations.

The datastructure side of it is breaking up n words into bits and interleaving them within n variables (usually machine registers), so that the first variable contains the first bit from each word, the second variable the second bit, etc. It’s also called “SIMD within a register”.