HTTP endpoints
For a list of API endpoints supported by the Coordinator, see .
Overview
The Druid Coordinator process is primarily responsible for segment management and distribution. More specifically, the Druid Coordinator process communicates to Historical processes to load or drop segments based on configurations. The Druid Coordinator is responsible for loading new segments, dropping outdated segments, ensuring that segments are “replicated” (that is, loaded on multiple different Historical nodes) proper (configured) number of times, and moving (“balancing”) segments between Historical nodes to keep the latter evenly loaded.
The Druid Coordinator runs its duties periodically and the time between each run is a configurable parameter. On each run, the Coordinator assesses the current state of the cluster before deciding on the appropriate actions to take. Similar to the Broker and Historical processes, the Druid Coordinator maintains a connection to a Zookeeper cluster for current cluster information. The Coordinator also maintains a connection to a database containing information about “used” segments (that is, the segments that should be loaded in the cluster) and the loading rules.
Before any unassigned segments are serviced by Historical processes, the Historical processes for each tier are first sorted in terms of capacity, with least capacity servers having the highest priority. Unassigned segments are always assigned to the processes with least capacity to maintain a level of balance between processes. The Coordinator does not directly communicate with a historical process when assigning it a new segment; instead the Coordinator creates some temporary information about the new segment under load queue path of the historical process. Once this request is seen, the historical process will load the segment and begin servicing it.
Running
Segments can be automatically loaded and dropped from the cluster based on a set of rules. For more information on rules, see Rule Configuration.
Cleaning up segments
On each run, the Druid Coordinator compares the set of used segments in the database with the segments served by some Historical nodes in the cluster. Coordinator sends requests to Historical nodes to unload unused segments or segments that are removed from the database.
Segments that are overshadowed (their versions are too old and their data has been replaced by newer segments) are marked as unused. During the next Coordinator’s run, they will be unloaded from Historical nodes in the cluster.
Segment availability
If a Historical process restarts or becomes unavailable for any reason, the Druid Coordinator will notice a process has gone missing and treat all segments served by that process as being dropped. Given a sufficient period of time, the segments may be reassigned to other Historical processes in the cluster. However, each segment that is dropped is not immediately forgotten. Instead, there is a transitional data structure that stores all dropped segments with an associated lifetime. The lifetime represents a period of time in which the Coordinator will not reassign a dropped segment. Hence, if a historical process becomes unavailable and available again within a short period of time, the historical process will start up and serve segments from its cache without any those segments being reassigned across the cluster.
Balancing segment load
Each run, the Druid Coordinator compacts segments by merging small segments or splitting a large one. This is useful when your segments are not optimized in terms of segment size which may degrade query performance. See Segment Size Optimization for details.
The Coordinator first finds the segments to compact based on the . Once some segments are found, it issues a compaction task to compact those segments. The maximum number of running compaction tasks is . Note that even though min(sum of worker capacity * slotRatio, maxSlots)
= 0, at least one compaction task is always submitted if the compaction is enabled for a dataSource. See and Compaction Configuration to enable the compaction.
Compaction tasks might fail due to the following reasons.
- If the input segments of a compaction task are removed or overshadowed before it starts, that compaction task fails immediately.
- If a task of a higher priority acquires a for an interval overlapping with the interval of a compaction task, the compaction task fails.
Once a compaction task fails, the Coordinator simply checks the segments in the interval of the failed task again, and issues another compaction task in the next run.
Segment search policy
Recent segment first policy
At every coordinator run, this policy looks up time chunks in order of newest-to-oldest and checks whether the segments in those time chunks need compaction or not. A set of segments need compaction if all conditions below are satisfied.
- Total size of segments in the time chunk is smaller than or equal to the configured
inputSegmentSizeBytes
.
Here are some details with an example. Suppose we have two dataSources (foo
, ) as seen below:
foo
foo_2017-11-01T00:00:00.000Z_2017-12-01T00:00:00.000Z_VERSION
foo_2017-11-01T00:00:00.000Z_2017-12-01T00:00:00.000Z_VERSION_1
foo_2017-09-01T00:00:00.000Z_2017-10-01T00:00:00.000Z_VERSION
bar
bar_2017-10-01T00:00:00.000Z_2017-11-01T00:00:00.000Z_VERSION
Assuming that each segment is 10 MB and haven’t been compacted yet, this policy first returns two segments of foo_2017-11-01T00:00:00.000Z_2017-12-01T00:00:00.000Z_VERSION
and foo_2017-11-01T00:00:00.000Z_2017-12-01T00:00:00.000Z_VERSION_1
to compact together because 2017-11-01T00:00:00.000Z/2017-12-01T00:00:00.000Z
is the most recent time chunk.
If the coordinator has enough task slots for compaction, this policy will continue searching for the next segments and return bar_2017-10-01T00:00:00.000Z_2017-11-01T00:00:00.000Z_VERSION
and bar_2017-10-01T00:00:00.000Z_2017-11-01T00:00:00.000Z_VERSION_1
. Finally, foo_2017-09-01T00:00:00.000Z_2017-10-01T00:00:00.000Z_VERSION
will be picked up even though there is only one segment in the time chunk of .
The Coordinator console
The Druid Coordinator exposes a web GUI for displaying cluster information and rule configuration. For more details, please see .
FAQ
Does it matter if the Coordinator process starts up before or after other processes?
No. If the Druid Coordinator is not started up, no new segments will be loaded in the cluster and outdated segments will not be dropped. However, the Coordinator process can be started up at any time, and after a configurable delay, will start running Coordinator tasks.