Skip to content

RagnarGrootKoerkamp/PtrHash

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

PtrHash: Minimal Perfect Hashing at RAM Throughput

crates.io docs.rs

PtrHash is a fast and space efficient minimal perfect hash function that maps a list of n distinct keys into {0,...,n-1}. It is based on/inspired by PTHash (and much more than just a Rust rewrite).

Read the preprint (arXiv, blog version) for details on the algorithm and benchmarks against other methods:

Ragnar Groot Koerkamp. PtrHash: Minimal Perfect Hashing at RAM Throughput. arXiv (2025). doi.org/10.48550/arXiv.2502.15539

Source code for the paper evals can be found in examples/evals.rs, and analysis is evals.py. Plots can be found in the blog.

Contact

In case you run into any kind of issue or things are unclear, please make issues and/or PRs, or reach out on twitter/bsky. I'm more than happy to help out with integrating PtrHash.

Performance

PtrHash supports up to 2^40 keys. For default parameters, constructing a MPHF of n=10^9 integer keys gives:

  • Construction takes 30s on my i7-10750H (2.6GHz) on 6 threads.
    • 6s to sort hashes,
    • 23s to find pilots.
  • Memory usage is 2.41bits/key:
    • 2.29bits/key for pilots,
    • 0.12bits/key for remapping.
  • Queries take:
    • 21ns/key when indexing sequentially,
    • 8.7ns/key when streaming with prefetching,
    • 2.6ns/key when streaming with prefetching, using 4 threads.
  • When giving up on minimality of the hash and allowing values up to n/alpha, query times slightly improve:
    • 17.6ns/key when indexing sequentially,
    • 7.9ns/key when streaming using prefetching,
    • 2.6ns/key when streaming with prefetching, using 4 threads.

Query throughput per thread fully saturates the prefetching bandwidth of each core, and multithreaded querying fully saturates the DDR4 memory bandwidth.

Input

PtrHash is primarily intended to be used on large sets of keys, say of size at least 1 million. Nevertheless, it can also be used for sets as small as e.g. 10 keys. In this case, there will be a relatively large constant space overhead, and other methods may be smaller and/or faster. (PtrHash should still be fast, but the small probability of having to remap values may be slow compared to methods designed for small inputs.)

Usage

Below, we use PtrHashParams::default() for a reasonable tradeoff between size (2.4 bits/key) and speed. Slightly smaller size is possible using PtrHashParams::default_compact(), at the cost of significantly slower construction time (2x) and lowered reliability.

There is also PtrHashParams::default_fast(), which takes 25% more space but can be almost 2x faster when querying integer keys in tight loops. Nevertheless, for large inputs, maximum query throughput is achieved with index_stream with default parameters.

use ptr_hash::{PtrHash, PtrHashParams};

// Generate some random keys.
let n = 1_000_000_000;
let keys = ptr_hash::util::generate_keys(n);

// Build the datastructure.
let mphf = <PtrHash>::new(&keys, PtrHashParams::default());

// Get the minimal index of a key.
let key = 0;
let idx = mphf.index(&key);
assert!(idx < n);

// Get the non-minimal index of a key. Slightly faster, but can be >=n.
let _idx = mphf.index_no_remap(&key);

// An iterator over the indices of the keys.
// 32: number of iterations ahead to prefetch.
// true: remap to a minimal key in [0, n).
let indices = mphf.index_stream::<32, true, _>(&keys);
assert_eq!(indices.sum::<usize>(), (n * (n - 1)) / 2);

// Test that all items map to different indices
let mut taken = vec![false; n];
for key in keys {
    let idx = mphf.index(&key);
    assert!(!taken[idx]);
    taken[idx] = true;
}

Epserde

The PtrHash datastructure can be (de)serialized to/from disk using epserde when the epserde feature is set. This also allows convenient deserialization using mmap. See examples/epserde.rs for an example.

Sharding

In order to build PtrHash on large sets of keys that do not fit in ram, the keys can be sharded and constructed one shard at a time. See fn sharding() in examples/evals.rs for an example.

Compared to PTHash

PtrHash extends PTHash in a few ways:

  • 8-bit pilots: Instead of allowing pilots to take any integer value, we restrict them to [0, 256) and store them as Vec<u8> directly. This avoids the need for a compact or dictionary encoding.

  • Evicting: To get all pilots to be small, we use evictions, similar to cuckoo hashing: Whenever we cannot find a collision-free pilot for a bucket, we find the pilot with the fewest collisions and evict all colliding buckets, which are pushed on a queue after which they will search for a new pilot.

  • Partitioning: To speed up construction, we partition all keys/hashes into parts such that each part contains S=2^k slots. This significantly speeds up construction since all reads of the taken bitvector are now very local.

    This brings the benefit that the only global memory needed is to store the hashes for each part. The sorting, bucketing, and slot filling is per-part and needs comparatively little memory.

  • Remap encoding: We use the CachelineEF partitioned Elias-Fano encoding that stores chunks of 44 integers into a single cacheline. This takes ~30% more space for remapping, but replaces the three reads needed by (global) Elias-Fano encoding by a single read.