PtrHash is a fast and space efficient minimal perfect hash function that maps
a list of n
distinct keys into {0,...,n-1}
.
It is based on/inspired by PTHash (and much
more than just a Rust rewrite).
Read the preprint (arXiv, blog version) for details on the algorithm and benchmarks against other methods:
Ragnar Groot Koerkamp. PtrHash: Minimal Perfect Hashing at RAM Throughput. arXiv (2025). doi.org/10.48550/arXiv.2502.15539
Source code for the paper evals can be found in examples/evals.rs, and analysis is evals.py. Plots can be found in the blog.
In case you run into any kind of issue or things are unclear, please make issues and/or PRs, or reach out on twitter/bsky. I'm more than happy to help out with integrating PtrHash.
PtrHash supports up to 2^40
keys. For default parameters, constructing a MPHF of n=10^9
integer keys gives:
- Construction takes
30s
on myi7-10750H
(2.6GHz
) on 6 threads.6s
to sort hashes,23s
to find pilots.
- Memory usage is
2.41bits/key
:2.29bits/key
for pilots,0.12bits/key
for remapping.
- Queries take:
21ns/key
when indexing sequentially,8.7ns/key
when streaming with prefetching,2.6ns/key
when streaming with prefetching, using4
threads.
- When giving up on minimality of the hash and allowing values up to
n/alpha
, query times slightly improve:17.6ns/key
when indexing sequentially,7.9ns/key
when streaming using prefetching,2.6ns/key
when streaming with prefetching, using4
threads.
Query throughput per thread fully saturates the prefetching bandwidth of each core, and multithreaded querying fully saturates the DDR4 memory bandwidth.
PtrHash is primarily intended to be used on large sets of keys, say of size at least 1 million. Nevertheless, it can also be used for sets as small as e.g. 10 keys. In this case, there will be a relatively large constant space overhead, and other methods may be smaller and/or faster. (PtrHash should still be fast, but the small probability of having to remap values may be slow compared to methods designed for small inputs.)
Below, we use PtrHashParams::default()
for a reasonable tradeoff between size
(2.4 bits/key) and speed.
Slightly smaller size is possible using PtrHashParams::default_compact()
,
at the cost of significantly slower construction time (2x) and lowered reliability.
There is also PtrHashParams::default_fast()
, which takes 25% more space but
can be almost 2x faster when querying integer keys in tight loops. Nevertheless,
for large inputs, maximum query throughput is achieved with index_stream
with default parameters.
use ptr_hash::{PtrHash, PtrHashParams};
// Generate some random keys.
let n = 1_000_000_000;
let keys = ptr_hash::util::generate_keys(n);
// Build the datastructure.
let mphf = <PtrHash>::new(&keys, PtrHashParams::default());
// Get the minimal index of a key.
let key = 0;
let idx = mphf.index(&key);
assert!(idx < n);
// Get the non-minimal index of a key. Slightly faster, but can be >=n.
let _idx = mphf.index_no_remap(&key);
// An iterator over the indices of the keys.
// 32: number of iterations ahead to prefetch.
// true: remap to a minimal key in [0, n).
let indices = mphf.index_stream::<32, true, _>(&keys);
assert_eq!(indices.sum::<usize>(), (n * (n - 1)) / 2);
// Test that all items map to different indices
let mut taken = vec![false; n];
for key in keys {
let idx = mphf.index(&key);
assert!(!taken[idx]);
taken[idx] = true;
}
The PtrHash
datastructure can be (de)serialized to/from disk using
epserde when the epserde
feature is set.
This also allows convenient deserialization using mmap
.
See examples/epserde.rs for an example.
In order to build PtrHash on large sets of keys that do not fit in ram, the keys
can be sharded and constructed one shard at a time.
See fn sharding()
in examples/evals.rs for an example.
PtrHash extends PTHash in a few ways:
-
8-bit pilots: Instead of allowing pilots to take any integer value, we restrict them to
[0, 256)
and store them asVec<u8>
directly. This avoids the need for a compact or dictionary encoding. -
Evicting: To get all pilots to be small, we use evictions, similar to cuckoo hashing: Whenever we cannot find a collision-free pilot for a bucket, we find the pilot with the fewest collisions and evict all colliding buckets, which are pushed on a queue after which they will search for a new pilot.
-
Partitioning: To speed up construction, we partition all keys/hashes into parts such that each part contains
S=2^k
slots. This significantly speeds up construction since all reads of thetaken
bitvector are now very local.This brings the benefit that the only global memory needed is to store the hashes for each part. The sorting, bucketing, and slot filling is per-part and needs comparatively little memory.
-
Remap encoding: We use the
CachelineEF
partitioned Elias-Fano encoding that stores chunks of44
integers into a single cacheline. This takes~30%
more space for remapping, but replaces the three reads needed by (global) Elias-Fano encoding by a single read.