I’m wondering if there is any tuning I should be doing to get the highest possible read performance? The seeks to the right key prefix seem to take a lot of time.
@gargan would you be able to share your data directory and a sample query that I can use to reproduce this? You can send the data directory to Ibrahim[at]dgraph.io .
Also, can you run the badger info --dir xxx command and share the output? This command runs on a inactive badger directory so you might have to stop the running application.
Great. Here is the output of the badger info command. The data to reproduce the query is about 2.3GB at the moment (won’t go through email). I can share it if you have a place I can push it to.
@gargan Can you try running your queries once again? Your data might be in level 0 and the DB close forced it to level 1. The level 0 to level 1 compaction would’ve dropped all the stale keys which should improve the read performance.
I realized I wasn’t doing an even comparison; there was some hidden caching going on with the file side. So I added the same caching on the badger side and now its only about 3x slower.
Still I’m trying to make reads go as fast as possible. Would these or any other options be expected to help?