After using the latest version, both the client and ratelUI are very slow queries. Is it because I query when I input data? If so, how do we need to solve it? Distributed?
If it is suspended, the log will tell me that 2 more transactions are being executed. What should I do?
I tested it according to your method, and the query would not be too slow. But there is a new problem, when I insert data, will occupy a lot of memory, resulting in my insert slower and slower, other operations can not be completed, the computer can not operate.
I used the way of inserting RDF to insert data, and also tried the way of JSON, and I had the same problem.
If you do not want to increase memory, I recommend using Bulkload for data insertion.
And which of the problems presented still persist?
A bulkload log looks like this:
./dgraph bulk -r out.rdf.gz -s movielens.schema --map_shards=4 --reduce_shards=1 --zero localhost:5080
{
"RDFDir": "out.rdf.gz",
"SchemaFile": "movielens.schema",
"DgraphsDir": "out",
"TmpDir": "tmp",
"NumGoroutines": 8,
"MapBufSize": 67108864,
"ExpandEdges": true,
"SkipMapPhase": false,
"CleanupTmp": true,
"NumShufflers": 1,
"Version": false,
"StoreXids": false,
"ZeroAddr": "localhost:5080",
"HttpAddr": "localhost:8080",
"MapShards": 4,
"ReduceShards": 1
}
The bulk loader needs to open many files at once. This number depends on the size of the data
set loaded, the map file output size, and the level of indexing. 100,000 is adequate for most data
set sizes. See `man ulimit` for details of how to change the limit.
Current max open files limit: 7168
2018/07/11 21:55:13 loader.go:77: Connecting to zero at localhost:5080
MAP 01s rdf_count:30.54k rdf_speed:29.83k/sec edge_count:94.38k edge_speed:92.17k/sec
MAP 02s rdf_count:108.4k rdf_speed:53.44k/sec edge_count:327.9k edge_speed:161.7k/sec
2018/07/11 21:55:15 merge_shards.go:36: Shard tmp/shards/002 -> Reduce tmp/shards/shard_0/002
2018/07/11 21:55:15 merge_shards.go:36: Shard tmp/shards/001 -> Reduce tmp/shards/shard_0/001
2018/07/11 21:55:15 merge_shards.go:36: Shard tmp/shards/000 -> Reduce tmp/shards/shard_0/000
2018/07/11 21:55:15 merge_shards.go:36: Shard tmp/shards/003 -> Reduce tmp/shards/shard_0/003
REDUCE 03s [100.00%] edge_count:327.9k edge_speed:327.9k/sec plist_count:16.63k plist_speed:16.63k/sec
REDUCE 03s [100.00%] edge_count:327.9k edge_speed:877.8k/sec plist_count:16.63k plist_speed:44.52k/sec
Total: 03s
I have load the data, but I don’t know how to do the next step.
And,can bulkloader only be used to initialize the database? What if I need to input data all the time? How does this method need to be used?
It is done just for this, to initialize an ingest from scratch.
In that case either you use a client or via LiveLoad.
is simple, BulkLoad creates the output folder “Out/*” each new folder with a number would be a Shard. You can control this by editing the flag “--reduce_shards”. Once you have a output done, just get those files in “out/0/*” and move to a Dgraph Server from scratch. And run your server.
Or copy the Dgraph binarie to that path and start Dgraph Server from there.
Yes.I did.
However, what I need is to insert data after starting the server.Because insertions and queries are possible at the same time.So,this method is not suitable,but other methods will occupied memory.
Is it possible that you record a video and send me in personal inbox(discuss)? showing exactly what you are doing by reproducing the issue. Doing it so I could easily determine an exit for you. (youtube or something accessible).
Are you using docker? At first print I see an unusual IP: “10.10.18.71”.
Another detail.
Its specs say it has 8GB of memory. Is not that too small for your operating system? possibly + docker + some other software + browser and etc?
In practice you have very few RDFs (Supposedly 12k, that’s very small, should not be a problem). 2GB for the Dgraph is more than enough. So there’s something you’re doing that does not fit.