Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

clarify README #23

Open
matu3ba opened this issue Mar 4, 2021 · 0 comments
Open

clarify README #23

matu3ba opened this issue Mar 4, 2021 · 0 comments

Comments

@matu3ba
Copy link

matu3ba commented Mar 4, 2021

RAM has >100ns memory access delay, whereas cache has lower than 60ns. This is abit vague in the README.
On top of that comes the hardware prefetcher that detects linear access patterns and has some dark magic for speculative execution with speculative cache prefetching.
Usually Linux memory-maps smaller files directly to L3 cache (when available), which is also not exactly formulated in the README.

Since the algorithm used is a streaming one with sliding window for vectorization: Is there a specific reason, why you do not use circular buffers as fastest available data structure?

Otherwise, it would be helpful to make a separate benchmark on allocator implementations only for reading files.

@matu3ba matu3ba changed the title clarify RAM <-> cache behavior clarify README Mar 4, 2021
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant