At present, the cluster has been running successfully. Now I use golang to write data to dgraph. From the official tutorial: https://github.com/dgraph-io/dgo, conn, err := grpc.Dial("localhost:9080", grpc.WithInsecure())
When I write data to dgraph I must specify address and port . In this way, isn’t all data written to the same node? So how do I ensure that the other three nodes in the cluster also have data?
If you’re not using replicas, the other Alpha instances will remain idle until the Dgraph Cluster recognizes the need to utilize them. This will be dependent on disk usage. However, you can employ the API to manually move predicates to other Alpha instances. Doing so will enable you to achieve the desired result.
When you have replicas or when you manually partition predicates by moving them to other Alphas, the Dgraph Cluster takes care of properly distributing the data. There’s no need to worry about which Alpha to send the mutation to.
Nonetheless, I would recommend implementing a “Round-robin scheduling” logic. It’s beneficial to maintain this practice even when starting small, in case your application scales up in the future. Round-robin, in this context, would help alleviate the load on a single Alpha instance. If you send all mutations to one place, it could overwhelm that instance alone. Therefore, load balancing is of paramount importance. This is the approach we take in Liveloader.
Does it mean that the data can be guaranteed to be backed up on each node? Then all the pressure is on the zero node. If I adopt the Round-robin scheduling method you mentioned,which means that the program needs to randomly connect to the dgraph client when writing data. Is there any official recommended method?
Not exactly. The zero has just a few jobs. Like timestamps, UID leasing, data balancing and so on. So it won’t be in pressure.
Not that I remember. But this is just an optimization. In general queries are distributed by nature. But the deserialization happens on a single node. And if you have an app with many clients, many requests. This could be a temperature point. So it’s better to distribute. To lessen the burden on the deserialization.
Thank you a lot! And I have a more comprehensive understanding of dgraph. In addition, if you have official channels in Asia? As far as I know, there is a great demand for graph databases in Asia, but communication is not very convenient due to the time difference between us.