new HashMap<ResourceId, ContainerProvider>();
@Override
public void process(ClusterEvent event) throws Exception {
final HelixManager helixManager = event.getAttribute("helixmanager");
final HelixAdmin helixAdmin = helixManager.getClusterManagmentTool();
final Map<ResourceId, ResourceConfig> resourceMap =
event.getAttribute(AttributeName.RESOURCES.toString());
final HelixDataAccessor accessor = helixManager.getHelixDataAccessor();
final PropertyKey.Builder keyBuilder = accessor.keyBuilder();
for (ResourceId resourceId : resourceMap.keySet()) {
ResourceConfig resourceConfig = resourceMap.get(resourceId);
ProvisionerConfig provisionerConfig = resourceConfig.getProvisionerConfig();
if (provisionerConfig != null) {
Provisioner provisioner;
provisioner = _provisionerMap.get(resourceId);
// instantiate and cache a provisioner if there isn't one already cached
if (provisioner == null) {
ProvisionerRef provisionerRef = provisionerConfig.getProvisionerRef();
if (provisionerRef != null) {
provisioner = provisionerRef.getProvisioner();
}
if (provisioner != null) {
provisioner.init(helixManager, resourceConfig);
_containerProviderMap.put(resourceId, provisioner.getContainerProvider());
_targetProviderMap.put(resourceId, provisioner.getTargetProvider());
_provisionerMap.put(resourceId, provisioner);
} else {
LOG.error("Resource " + resourceId + " does not have a valid provisioner class!");
break;
}
}
TargetProvider targetProvider = _targetProviderMap.get(resourceId);
ContainerProvider containerProvider = _containerProviderMap.get(resourceId);
final Cluster cluster = event.getAttribute("Cluster");
final ClusterDataCache cache = event.getAttribute("ClusterDataCache");
final Collection<Participant> participants = cluster.getParticipantMap().values();
// If a process died, we need to mark it as DISCONNECTED or if the process is ready, mark as
// CONNECTED
Map<ParticipantId, Participant> participantMap = cluster.getParticipantMap();
for (ParticipantId participantId : participantMap.keySet()) {
Participant participant = participantMap.get(participantId);
ContainerConfig config = participant.getContainerConfig();
if (config != null) {
ContainerState containerState = config.getState();
if (!participant.isAlive() && ContainerState.CONNECTED.equals(containerState)) {
// Need to mark as disconnected if process died
LOG.info("Participant " + participantId + " died, marking as DISCONNECTED");
updateContainerState(cache, accessor, keyBuilder, cluster, null, participantId,
ContainerState.DISCONNECTED);
} else if (participant.isAlive() && ContainerState.CONNECTING.equals(containerState)) {
// Need to mark as connected only when the live instance is visible
LOG.info("Participant " + participantId + " is ready, marking as CONNECTED");
updateContainerState(cache, accessor, keyBuilder, cluster, null, participantId,
ContainerState.CONNECTED);
} else if (!participant.isAlive() && ContainerState.HALTING.equals(containerState)) {
// Need to mark as connected only when the live instance is visible
LOG.info("Participant " + participantId + " is has been killed, marking as HALTED");
updateContainerState(cache, accessor, keyBuilder, cluster, null, participantId,
ContainerState.HALTED);
}
}
}
// Participants registered in helix
// Give those participants to targetprovider
// Provide the response that contains, new containerspecs, containers to be released,
// containers to be stopped
// call the respective provisioner to allocate and start the container.
// Each container is then started its state is changed from any place.
// The target provider is given the state of container and asked for its new state. For each
// state there is a corresponding handler function.
// TargetProvider should be stateless, given the state of cluster and existing participants
// it should return the same result
final TargetProviderResponse response =
targetProvider.evaluateExistingContainers(cluster, resourceId, participants);
// allocate new containers
for (final ContainerSpec spec : response.getContainersToAcquire()) {
final ParticipantId participantId = spec.getParticipantId();
if (!cluster.getParticipantMap().containsKey(participantId)) {
// create a new Participant, attach the container spec
InstanceConfig instanceConfig = new InstanceConfig(participantId);
instanceConfig.setInstanceEnabled(false);
instanceConfig.setContainerSpec(spec);
// create a helix_participant in ACQUIRING state
instanceConfig.setContainerState(ContainerState.ACQUIRING);
// create the helix participant and add it to cluster
helixAdmin.addInstance(cluster.getId().toString(), instanceConfig);
cache.requireFullRefresh();
}
LOG.info("Allocating container for " + participantId);
ListenableFuture<ContainerId> future = containerProvider.allocateContainer(spec);
FutureCallback<ContainerId> callback = new FutureCallback<ContainerId>() {
@Override
public void onSuccess(ContainerId containerId) {
LOG.info("Container " + containerId + " acquired. Marking " + participantId);
updateContainerState(cache, accessor, keyBuilder, cluster, containerId,
participantId, ContainerState.ACQUIRED);
}
@Override
public void onFailure(Throwable t) {
LOG.error("Could not allocate a container for participant " + participantId, t);
updateContainerState(cache, accessor, keyBuilder, cluster, null, participantId,
ContainerState.FAILED);
}
};
safeAddCallback(future, callback);
}
// start new containers
for (final Participant participant : response.getContainersToStart()) {
final ContainerId containerId = participant.getInstanceConfig().getContainerId();
updateContainerState(cache, accessor, keyBuilder, cluster, null, participant.getId(),
ContainerState.CONNECTING);
// create the helix participant and add it to cluster
LOG.info("Starting container " + containerId + " for " + participant.getId());
ListenableFuture<Boolean> future =
containerProvider.startContainer(containerId, participant);
FutureCallback<Boolean> callback = new FutureCallback<Boolean>() {
@Override
public void onSuccess(Boolean result) {
// Do nothing yet, need to wait for live instance
LOG.info("Container " + containerId + " started for " + participant.getId());
}
@Override
public void onFailure(Throwable t) {
LOG.error("Could not start container" + containerId + "for participant "
+ participant.getId(), t);
updateContainerState(cache, accessor, keyBuilder, cluster, null, participant.getId(),
ContainerState.FAILED);
}
};
safeAddCallback(future, callback);
}
// release containers
for (final Participant participant : response.getContainersToRelease()) {
// mark it as finalizing
final ContainerId containerId = participant.getInstanceConfig().getContainerId();
updateContainerState(cache, accessor, keyBuilder, cluster, null, participant.getId(),
ContainerState.FINALIZING);
// remove the participant
LOG.info("Deallocating container " + containerId + " for " + participant.getId());
ListenableFuture<Boolean> future = containerProvider.deallocateContainer(containerId);
FutureCallback<Boolean> callback = new FutureCallback<Boolean>() {
@Override
public void onSuccess(Boolean result) {
LOG.info("Container " + containerId + " deallocated. Dropping " + participant.getId());
InstanceConfig existingInstance =
helixAdmin.getInstanceConfig(cluster.getId().toString(), participant.getId()
.toString());
helixAdmin.dropInstance(cluster.getId().toString(), existingInstance);
cache.requireFullRefresh();
}
@Override
public void onFailure(Throwable t) {