/*
|
* CDDL HEADER START
|
*
|
* The contents of this file are subject to the terms of the
|
* Common Development and Distribution License, Version 1.0 only
|
* (the "License"). You may not use this file except in compliance
|
* with the License.
|
*
|
* You can obtain a copy of the license at
|
* trunk/opends/resource/legal-notices/OpenDS.LICENSE
|
* or https://OpenDS.dev.java.net/OpenDS.LICENSE.
|
* See the License for the specific language governing permissions
|
* and limitations under the License.
|
*
|
* When distributing Covered Code, include this CDDL HEADER in each
|
* file and include the License file at
|
* trunk/opends/resource/legal-notices/OpenDS.LICENSE. If applicable,
|
* add the following below this CDDL HEADER, with the fields enclosed
|
* by brackets "[]" replaced with your own identifying information:
|
* Portions Copyright [yyyy] [name of copyright owner]
|
*
|
* CDDL HEADER END
|
*
|
*
|
* Copyright 2008-2010 Sun Microsystems, Inc.
|
* Portions Copyright 2011-2013 ForgeRock AS
|
*/
|
package org.opends.server.replication.service;
|
|
import java.io.BufferedOutputStream;
|
import java.io.IOException;
|
import java.io.InputStream;
|
import java.io.OutputStream;
|
import java.net.SocketTimeoutException;
|
import java.util.*;
|
import java.util.concurrent.ConcurrentHashMap;
|
import java.util.concurrent.TimeoutException;
|
import java.util.concurrent.atomic.AtomicBoolean;
|
import java.util.concurrent.atomic.AtomicInteger;
|
|
import org.opends.messages.Category;
|
import org.opends.messages.Message;
|
import org.opends.messages.Severity;
|
import org.opends.server.api.DirectoryThread;
|
import org.opends.server.backends.task.Task;
|
import org.opends.server.config.ConfigException;
|
import org.opends.server.loggers.debug.DebugTracer;
|
import org.opends.server.replication.common.*;
|
import org.opends.server.replication.protocol.*;
|
import org.opends.server.tasks.InitializeTargetTask;
|
import org.opends.server.tasks.InitializeTask;
|
import org.opends.server.types.Attribute;
|
import org.opends.server.types.DN;
|
import org.opends.server.types.DirectoryException;
|
import org.opends.server.types.ResultCode;
|
|
import static org.opends.messages.ReplicationMessages.*;
|
import static org.opends.server.loggers.ErrorLogger.*;
|
import static org.opends.server.loggers.debug.DebugLogger.*;
|
import static org.opends.server.replication.common.StatusMachine.*;
|
|
/**
|
* This class should be used as a base for Replication implementations.
|
* <p>
|
* It is intended that developer in need of a replication mechanism
|
* subclass this class with their own implementation.
|
* <p>
|
* The startup phase of the ReplicationDomain subclass,
|
* should read the list of replication servers from the configuration,
|
* instantiate a {@link ServerState} then start the publish service
|
* by calling
|
* {@link #startPublishService(Set, int, long, long)}.
|
* At this point it can start calling the {@link #publish(UpdateMsg)}
|
* method if needed.
|
* <p>
|
* When the startup phase reach the point when the subclass is ready
|
* to handle updates the Replication Domain implementation should call the
|
* {@link #startListenService()} method.
|
* At this point a Listener thread is created on the Replication Service
|
* and which can start receiving updates.
|
* <p>
|
* When updates are received the Replication Service calls the
|
* {@link #processUpdate(UpdateMsg, AtomicBoolean)} method.
|
* ReplicationDomain implementation should implement the appropriate code
|
* for replaying the update on the local repository.
|
* When fully done the subclass must call the
|
* {@link #processUpdateDone(UpdateMsg, String)} method.
|
* This allows to process the update asynchronously if necessary.
|
*
|
* <p>
|
* To propagate changes to other replica, a ReplicationDomain implementation
|
* must use the {@link #publish(UpdateMsg)} method.
|
* <p>
|
* If the Full Initialization process is needed then implementation
|
* for {@code importBackend(InputStream)} and
|
* {@code exportBackend(OutputStream)} must be
|
* provided.
|
* <p>
|
* Full Initialization of a replica can be triggered by LDAP clients
|
* by creating InitializeTasks or InitializeTargetTask.
|
* Full initialization can also by triggered from the ReplicationDomain
|
* implementation using methods {@link #initializeRemote(int)}
|
* or {@link #initializeFromRemote(int)}.
|
* <p>
|
* At shutdown time, the {@link #stopDomain()} method should be called to
|
* cleanly stop the replication service.
|
*/
|
public abstract class ReplicationDomain
|
{
|
/**
|
* Current status for this replicated domain.
|
*/
|
protected ServerStatus status = ServerStatus.NOT_CONNECTED_STATUS;
|
|
/**
|
* The tracer object for the debug logger.
|
*/
|
private static final DebugTracer TRACER = getTracer();
|
|
/**
|
* The baseDN for the Replication Service.
|
* All Replication Domain using this baseDN will be connected
|
* through the Replication Service.
|
*/
|
private final DN baseDN;
|
|
/**
|
* The identifier of this Replication Domain inside the
|
* Replication Service.
|
* Each Domain must use a unique ServerID.
|
*/
|
private final int serverID;
|
|
/**
|
* The ReplicationBroker that is used by this ReplicationDomain to
|
* connect to the ReplicationService.
|
*/
|
protected ReplicationBroker broker = null;
|
|
/**
|
* This Map is used to store all outgoing assured messages in order
|
* to be able to correlate all the coming back acks to the original
|
* operation.
|
*/
|
private final Map<CSN, UpdateMsg> waitingAckMsgs =
|
new ConcurrentHashMap<CSN, UpdateMsg>();
|
|
|
/**
|
* The context related to an import or export being processed
|
* Null when none is being processed.
|
*/
|
protected IEContext ieContext = null;
|
|
/**
|
* The Thread waiting for incoming update messages for this domain and pushing
|
* them to the global incoming update message queue for later processing by
|
* replay threads.
|
*/
|
private ListenerThread listenerThread;
|
|
/**
|
* A Map used to store all the ReplicationDomains created on this server.
|
*/
|
private static Map<DN, ReplicationDomain> domains =
|
new HashMap<DN, ReplicationDomain>();
|
|
/*
|
* Assured mode properties
|
*/
|
/** Whether assured mode is enabled for this domain. */
|
private boolean assured = false;
|
/** Assured sub mode (used when assured is true). */
|
private AssuredMode assuredMode = AssuredMode.SAFE_DATA_MODE;
|
/** Safe Data level (used when assuredMode is SAFE_DATA). */
|
private byte assuredSdLevel = 1;
|
/** The timeout in ms that should be used, when waiting for assured acks. */
|
private long assuredTimeout = 2000;
|
|
/** Group id. */
|
private byte groupId = 1;
|
/**
|
* Referrals urls to be published to other servers of the topology.
|
* <p>
|
* TODO: fill that with all currently opened urls if no urls configured
|
*/
|
private final List<String> refUrls = new ArrayList<String>();
|
|
/**
|
* A set of counters used for Monitoring.
|
*/
|
private AtomicInteger numProcessedUpdates = new AtomicInteger(0);
|
private AtomicInteger numRcvdUpdates = new AtomicInteger(0);
|
private AtomicInteger numSentUpdates = new AtomicInteger(0);
|
|
/** Assured replication monitoring counters. */
|
|
/** Number of updates sent in Assured Mode, Safe Read. */
|
private AtomicInteger assuredSrSentUpdates = new AtomicInteger(0);
|
/**
|
* Number of updates sent in Assured Mode, Safe Read, that have been
|
* successfully acknowledged.
|
*/
|
private AtomicInteger assuredSrAcknowledgedUpdates = new AtomicInteger(0);
|
/**
|
* Number of updates sent in Assured Mode, Safe Read, that have not been
|
* successfully acknowledged (either because of timeout, wrong status or error
|
* at replay).
|
*/
|
private AtomicInteger assuredSrNotAcknowledgedUpdates =
|
new AtomicInteger(0);
|
/**
|
* Number of updates sent in Assured Mode, Safe Read, that have not been
|
* successfully acknowledged because of timeout.
|
*/
|
private AtomicInteger assuredSrTimeoutUpdates = new AtomicInteger(0);
|
/**
|
* Number of updates sent in Assured Mode, Safe Read, that have not been
|
* successfully acknowledged because of wrong status.
|
*/
|
private AtomicInteger assuredSrWrongStatusUpdates = new AtomicInteger(0);
|
/**
|
* Number of updates sent in Assured Mode, Safe Read, that have not been
|
* successfully acknowledged because of replay error.
|
*/
|
private AtomicInteger assuredSrReplayErrorUpdates = new AtomicInteger(0);
|
/**
|
* Multiple values allowed: number of updates sent in Assured Mode, Safe Read,
|
* that have not been successfully acknowledged (either because of timeout,
|
* wrong status or error at replay) for a particular server (DS or RS).
|
* <p>
|
* String format: <server id>:<number of failed updates>
|
*/
|
private final Map<Integer, Integer> assuredSrServerNotAcknowledgedUpdates =
|
new HashMap<Integer,Integer>();
|
/** Number of updates received in Assured Mode, Safe Read request. */
|
private AtomicInteger assuredSrReceivedUpdates = new AtomicInteger(0);
|
/**
|
* Number of updates received in Assured Mode, Safe Read request that we have
|
* acked without errors.
|
*/
|
private AtomicInteger assuredSrReceivedUpdatesAcked = new AtomicInteger(0);
|
/**
|
* Number of updates received in Assured Mode, Safe Read request that we have
|
* acked with errors.
|
*/
|
private AtomicInteger assuredSrReceivedUpdatesNotAcked = new AtomicInteger(0);
|
/** Number of updates sent in Assured Mode, Safe Data. */
|
private AtomicInteger assuredSdSentUpdates = new AtomicInteger(0);
|
/**
|
* Number of updates sent in Assured Mode, Safe Data, that have been
|
* successfully acknowledged.
|
*/
|
private AtomicInteger assuredSdAcknowledgedUpdates = new AtomicInteger(0);
|
/**
|
* Number of updates sent in Assured Mode, Safe Data, that have not been
|
* successfully acknowledged because of timeout.
|
*/
|
private AtomicInteger assuredSdTimeoutUpdates = new AtomicInteger(0);
|
/**
|
* Multiple values allowed: number of updates sent in Assured Mode, Safe Data,
|
* that have not been successfully acknowledged because of timeout for a
|
* particular RS.
|
* <p>
|
* String format: <server id>:<number of failed updates>
|
*/
|
private final Map<Integer, Integer> assuredSdServerTimeoutUpdates =
|
new HashMap<Integer,Integer>();
|
|
/**
|
* Window size used during initialization .. between
|
* - the initializer/exporter DS that listens/waits acknowledges and that
|
* slows down data msg publishing based on the slowest server
|
* - and each initialized/importer DS that publishes acknowledges each
|
* WINDOW/2 data msg received.
|
*/
|
protected int initWindow = 100;
|
|
/* Status related monitoring fields */
|
|
/**
|
* Indicates the date when the status changed. This may be used to indicate
|
* the date the session with the current replication server started (when
|
* status is NORMAL for instance). All the above assured monitoring fields
|
* are also reset each time the status is changed
|
*/
|
private Date lastStatusChangeDate = new Date();
|
|
/**
|
* The state maintained by the Concrete Class.
|
*/
|
private final ServerState state;
|
|
/**
|
* The generator that will be used to generate {@link CSN}
|
* for this domain.
|
*/
|
private final CSNGenerator generator;
|
|
private final Object eclIncludesLock = new Object();
|
private final Map<Integer, Set<String>> eclIncludesByServer =
|
new HashMap<Integer, Set<String>>();
|
private Set<String> eclIncludesAllServers = Collections.emptySet();
|
|
private final Map<Integer, Set<String>> eclIncludesForDeletesByServer =
|
new HashMap<Integer, Set<String>>();
|
private Set<String> eclIncludesForDeletesAllServers = Collections
|
.emptySet();
|
|
/**
|
* An object used to protect the initialization of the underlying broker
|
* session of this ReplicationDomain.
|
*/
|
private final Object sessionLock = new Object();
|
|
/**
|
* Returns the {@link CSNGenerator} that will be used to
|
* generate {@link CSN} for this domain.
|
*
|
* @return The {@link CSNGenerator} that will be used to
|
* generate {@link CSN} for this domain.
|
*/
|
public CSNGenerator getGenerator()
|
{
|
return generator;
|
}
|
|
/**
|
* Creates a ReplicationDomain with the provided parameters.
|
*
|
* @param baseDN The identifier of the Replication Domain to which
|
* this object is participating.
|
* @param serverID The identifier of the server that is participating
|
* to the Replication Domain.
|
* This identifier should be different for each server that
|
* is participating to a given Replication Domain.
|
* @param initWindow Window used during initialization.
|
*/
|
public ReplicationDomain(DN baseDN, int serverID, int initWindow)
|
{
|
this.baseDN = baseDN;
|
this.serverID = serverID;
|
this.initWindow = initWindow;
|
this.state = new ServerState();
|
this.generator = new CSNGenerator(serverID, state);
|
|
domains.put(baseDN, this);
|
}
|
|
/**
|
* Creates a ReplicationDomain with the provided parameters.
|
* (for unit test purpose only)
|
*
|
* @param baseDN The identifier of the Replication Domain to which
|
* this object is participating.
|
* @param serverID The identifier of the server that is participating
|
* to the Replication Domain.
|
* This identifier should be different for each server that
|
* is participating to a given Replication Domain.
|
* @param serverState The serverState to use
|
*/
|
public ReplicationDomain(DN baseDN, int serverID, ServerState serverState)
|
{
|
this.baseDN = baseDN;
|
this.serverID = serverID;
|
this.state = serverState;
|
this.generator = new CSNGenerator(serverID, state);
|
|
domains.put(baseDN, this);
|
}
|
|
/**
|
* Set the initial status of the domain and perform necessary initializations.
|
* This method will be called by the Broker each time the ReplicationBroker
|
* establish a new session to a Replication Server.
|
*
|
* Implementations may override this method when they need to perform
|
* additional computing after session establishment.
|
* The default implementation should be sufficient for ReplicationDomains
|
* that don't need to perform additional computing.
|
*
|
* @param initStatus The status to enter the state machine with.
|
* @param replicationServerState The ServerState of the ReplicationServer
|
* with which the session was established.
|
* @param generationID The current generationID of the
|
* ReplicationServer with which the session
|
* was established.
|
* @param session The Session that is currently used.
|
*/
|
public void sessionInitiated(
|
ServerStatus initStatus,
|
ServerState replicationServerState,
|
long generationID,
|
Session session)
|
{
|
// Sanity check: is it a valid initial status?
|
if (!isValidInitialStatus(initStatus))
|
{
|
Message msg = ERR_DS_INVALID_INIT_STATUS.get(initStatus.toString(),
|
getBaseDNString(), Integer.toString(serverID));
|
logError(msg);
|
} else
|
{
|
status = initStatus;
|
}
|
generator.adjust(state);
|
generator.adjust(replicationServerState);
|
}
|
|
/**
|
* Processes an incoming ChangeStatusMsg. Compute new status according to
|
* given order. Then update domain for being compliant with new status
|
* definition.
|
* @param csMsg The received status message
|
*/
|
private void receiveChangeStatus(ChangeStatusMsg csMsg)
|
{
|
if (debugEnabled())
|
TRACER.debugInfo("Replication domain " + baseDN +
|
" received change status message:\n" + csMsg);
|
|
ServerStatus reqStatus = csMsg.getRequestedStatus();
|
|
// Translate requested status to a state machine event
|
StatusMachineEvent event = StatusMachineEvent.statusToEvent(reqStatus);
|
if (event == StatusMachineEvent.INVALID_EVENT)
|
{
|
Message msg = ERR_DS_INVALID_REQUESTED_STATUS.get(reqStatus.toString(),
|
getBaseDNString(), Integer.toString(serverID));
|
logError(msg);
|
return;
|
}
|
|
// Set the new status to the requested one
|
setNewStatus(event);
|
}
|
|
/**
|
* Called when first connection or disconnection detected.
|
*/
|
void toNotConnectedStatus()
|
{
|
// Go into not connected status
|
setNewStatus(StatusMachineEvent.TO_NOT_CONNECTED_STATUS_EVENT);
|
}
|
|
/**
|
* Perform whatever actions are needed to apply properties for being
|
* compliant with new status. Must be called in synchronized section for
|
* status. The new status is already set in status variable.
|
*/
|
private void updateDomainForNewStatus()
|
{
|
switch (status)
|
{
|
case NOT_CONNECTED_STATUS:
|
break;
|
case NORMAL_STATUS:
|
break;
|
case DEGRADED_STATUS:
|
break;
|
case FULL_UPDATE_STATUS:
|
// Signal RS we just entered the full update status
|
broker.signalStatusChange(status);
|
break;
|
case BAD_GEN_ID_STATUS:
|
break;
|
default:
|
if (debugEnabled())
|
TRACER.debugInfo("updateDomainForNewStatus: unexpected status: " +
|
status);
|
}
|
}
|
|
/**
|
* Gets the status for this domain.
|
* @return The status for this domain.
|
*/
|
public ServerStatus getStatus()
|
{
|
return status;
|
}
|
|
/**
|
* Returns the base DN of this ReplicationDomain.
|
*
|
* @return The base DN of this ReplicationDomain
|
*/
|
public DN getBaseDN()
|
{
|
return baseDN;
|
}
|
|
/**
|
* Gets the baseDN of this domain.
|
*
|
* @return The baseDN for this domain.
|
*/
|
public String getBaseDNString()
|
{
|
return baseDN.toNormalizedString();
|
}
|
|
/**
|
* Get the server ID.
|
* @return The server ID.
|
*/
|
public int getServerId()
|
{
|
return serverID;
|
}
|
|
/**
|
* Tells if assured replication is enabled for this domain.
|
* @return True if assured replication is enabled for this domain.
|
*/
|
public boolean isAssured()
|
{
|
return assured;
|
}
|
|
/**
|
* Gives the mode for the assured replication of the domain.
|
* @return The mode for the assured replication of the domain.
|
*/
|
public AssuredMode getAssuredMode()
|
{
|
return assuredMode;
|
}
|
|
/**
|
* Gives the assured level of the replication of the domain.
|
* @return The assured level of the replication of the domain.
|
*/
|
public byte getAssuredSdLevel()
|
{
|
return assuredSdLevel;
|
}
|
|
/**
|
* Gives the assured timeout of the replication of the domain (in ms).
|
* @return The assured timeout of the replication of the domain.
|
*/
|
public long getAssuredTimeout()
|
{
|
return assuredTimeout;
|
}
|
|
/**
|
* Gets the group id for this domain.
|
* @return The group id for this domain.
|
*/
|
public byte getGroupId()
|
{
|
return groupId;
|
}
|
|
/**
|
* Gets the referrals URLs this domain publishes.
|
* @return The referrals URLs this domain publishes.
|
*/
|
public List<String> getRefUrls()
|
{
|
return refUrls;
|
}
|
|
/**
|
* Gets the info for Replicas in the topology (except us).
|
* @return The info for Replicas in the topology (except us)
|
*/
|
public List<DSInfo> getReplicasList()
|
{
|
return broker.getDsList();
|
}
|
|
/**
|
* Returns information about the DS server related to the provided serverId.
|
* based on the TopologyMsg we received when the remote replica connected or
|
* disconnected. Return null when no server with the provided serverId is
|
* connected.
|
*
|
* @param serverId The provided serverId of the remote replica
|
* @return the info related to this remote server if it is connected,
|
* null is the server is NOT connected.
|
*/
|
public DSInfo isRemoteDSConnected(int serverId)
|
{
|
for (DSInfo remoteDS : getReplicasList())
|
if (remoteDS.getDsId() == serverId)
|
return remoteDS;
|
return null;
|
}
|
|
/**
|
* Gets the States of all the Replicas currently in the
|
* Topology.
|
* When this method is called, a Monitoring message will be sent
|
* to the Replication Server to which this domain is currently connected
|
* so that it computes a table containing information about
|
* all Directory Servers in the topology.
|
* This Computation involves communications will all the servers
|
* currently connected and
|
*
|
* @return The States of all Replicas in the topology (except us)
|
*/
|
public Map<Integer, ServerState> getReplicaStates()
|
{
|
return broker.getReplicaStates();
|
}
|
|
/**
|
* Gets the info for RSs in the topology (except the one we are connected
|
* to).
|
* @return The info for RSs in the topology (except the one we are connected
|
* to)
|
*/
|
public List<RSInfo> getRsList()
|
{
|
return broker.getRsList();
|
}
|
|
|
/**
|
* Gets the server ID of the Replication Server to which the domain
|
* is currently connected.
|
*
|
* @return The server ID of the Replication Server to which the domain
|
* is currently connected.
|
*/
|
public int getRsServerId()
|
{
|
return broker.getRsServerId();
|
}
|
|
/**
|
* Increment the number of processed updates.
|
*/
|
private void incProcessedUpdates()
|
{
|
numProcessedUpdates.incrementAndGet();
|
}
|
|
/**
|
* get the number of updates replayed by the replication.
|
*
|
* @return The number of updates replayed by the replication
|
*/
|
int getNumProcessedUpdates()
|
{
|
if (numProcessedUpdates != null)
|
return numProcessedUpdates.get();
|
return 0;
|
}
|
|
/**
|
* get the number of updates received by the replication plugin.
|
*
|
* @return the number of updates received
|
*/
|
int getNumRcvdUpdates()
|
{
|
if (numRcvdUpdates != null)
|
return numRcvdUpdates.get();
|
return 0;
|
}
|
|
/**
|
* Get the number of updates sent by the replication plugin.
|
*
|
* @return the number of updates sent
|
*/
|
int getNumSentUpdates()
|
{
|
if (numSentUpdates != null)
|
return numSentUpdates.get();
|
return 0;
|
}
|
|
/**
|
* Set the list of Referrals that should be returned when an
|
* operation needs to be redirected to this server.
|
*
|
* @param referralsUrl The list of referrals.
|
*/
|
public void setURLs(Set<String> referralsUrl)
|
{
|
this.refUrls.addAll(referralsUrl);
|
}
|
|
/**
|
* Set the timeout of the assured replication.
|
*
|
* @param assuredTimeout the timeout of the assured replication.
|
*/
|
public void setAssuredTimeout(long assuredTimeout)
|
{
|
this.assuredTimeout = assuredTimeout;
|
}
|
|
/**
|
* Sets the groupID.
|
*
|
* @param groupId The groupID.
|
*/
|
public void setGroupId(byte groupId)
|
{
|
this.groupId = groupId;
|
}
|
|
/**
|
* Sets the level of assured replication.
|
*
|
* @param assuredSdLevel The level of assured replication.
|
*/
|
public void setAssuredSdLevel(byte assuredSdLevel)
|
{
|
this.assuredSdLevel = assuredSdLevel;
|
}
|
|
/**
|
* Sets the assured replication mode.
|
*
|
* @param dataMode The assured replication mode.
|
*/
|
public void setAssuredMode(AssuredMode dataMode)
|
{
|
this.assuredMode = dataMode;
|
}
|
|
/**
|
* Sets assured replication.
|
*
|
* @param assured A boolean indicating if assured replication should be used.
|
*/
|
public void setAssured(boolean assured)
|
{
|
this.assured = assured;
|
}
|
|
/**
|
* Receives an update message from the replicationServer.
|
* The other types of messages are processed in an opaque way for the caller.
|
* Also responsible for updating the list of pending changes
|
* @return the received message - null if none
|
*/
|
UpdateMsg receive()
|
{
|
UpdateMsg update = null;
|
|
while (update == null)
|
{
|
InitializeRequestMsg initReqMsg = null;
|
ReplicationMsg msg;
|
try
|
{
|
msg = broker.receive(true, true, false);
|
if (msg == null)
|
{
|
// The server is in the shutdown process
|
return null;
|
}
|
|
if (debugEnabled() && !(msg instanceof HeartbeatMsg))
|
{
|
TRACER.debugVerbose("Message received <" + msg + ">");
|
}
|
|
if (msg instanceof AckMsg)
|
{
|
AckMsg ack = (AckMsg) msg;
|
receiveAck(ack);
|
}
|
else if (msg instanceof InitializeRequestMsg)
|
{
|
// Another server requests us to provide entries
|
// for a total update
|
initReqMsg = (InitializeRequestMsg)msg;
|
}
|
else if (msg instanceof InitializeTargetMsg)
|
{
|
// Another server is exporting its entries to us
|
InitializeTargetMsg initTargetMsg = (InitializeTargetMsg) msg;
|
|
/*
|
This must be done while we are still holding the broker lock
|
because we are now going to receive a bunch of entries from the
|
remote server and we want the import thread to catch them and
|
not the ListenerThread.
|
*/
|
initialize(initTargetMsg, initTargetMsg.getSenderID());
|
}
|
else if (msg instanceof ErrorMsg)
|
{
|
ErrorMsg errorMsg = (ErrorMsg)msg;
|
if (ieContext != null)
|
{
|
/*
|
This is an error termination for the 2 following cases :
|
- either during an export
|
- or before an import really started
|
For example, when we publish a request and the
|
replicationServer did not find the import source.
|
|
A remote error during the import will be received in the
|
receiveEntryBytes() method.
|
*/
|
if (debugEnabled())
|
TRACER.debugInfo(
|
"[IE] processErrorMsg:" + this.serverID +
|
" baseDN: " + this.baseDN +
|
" Error Msg received: " + errorMsg);
|
|
if (errorMsg.getCreationTime() > ieContext.startTime)
|
{
|
// consider only ErrorMsg that relate to the current import/export
|
processErrorMsg(errorMsg);
|
}
|
else
|
{
|
/*
|
Simply log - happen when the ErrorMsg relates to a previous
|
attempt of initialization while we have started a new one
|
on this side.
|
*/
|
logError(ERR_ERROR_MSG_RECEIVED.get(errorMsg.getDetails()));
|
}
|
}
|
else
|
{
|
// Simply log - happen if import/export has been terminated
|
// on our side before receiving this ErrorMsg.
|
logError(ERR_ERROR_MSG_RECEIVED.get(errorMsg.getDetails()));
|
}
|
}
|
else if (msg instanceof ChangeStatusMsg)
|
{
|
ChangeStatusMsg csMsg = (ChangeStatusMsg)msg;
|
receiveChangeStatus(csMsg);
|
}
|
else if (msg instanceof UpdateMsg)
|
{
|
update = (UpdateMsg) msg;
|
generator.adjust(update.getCSN());
|
}
|
else if (msg instanceof InitializeRcvAckMsg)
|
{
|
if (ieContext != null)
|
{
|
InitializeRcvAckMsg ackMsg = (InitializeRcvAckMsg) msg;
|
ieContext.setAckVal(ackMsg.getSenderID(), ackMsg.getNumAck());
|
}
|
// Trash this msg When no input/export is running/should never happen
|
}
|
}
|
catch (SocketTimeoutException e)
|
{
|
// just retry
|
}
|
/*
|
Test if we have received and export request message and
|
if that's the case handle it now.
|
This must be done outside of the portion of code protected
|
by the broker lock so that we keep receiving update
|
when we are doing and export and so that a possible
|
closure of the socket happening when we are publishing the
|
entries to the remote can be handled by the other
|
replay thread when they call this method and therefore the
|
broker.receive() method.
|
*/
|
if (initReqMsg != null)
|
{
|
// Do this work in a thread to allow replay thread continue working
|
ExportThread exportThread = new ExportThread(
|
initReqMsg.getSenderID(), initReqMsg.getInitWindow());
|
exportThread.start();
|
}
|
}
|
|
numRcvdUpdates.incrementAndGet();
|
byte rsGroupId = broker.getRsGroupId();
|
if (update.isAssured()
|
&& update.getAssuredMode() == AssuredMode.SAFE_READ_MODE
|
&& rsGroupId == groupId)
|
{
|
assuredSrReceivedUpdates.incrementAndGet();
|
}
|
return update;
|
}
|
|
/**
|
* Updates the passed monitoring list of errors received for assured messages
|
* (safe data or safe read, depending of the passed list to update) for a
|
* particular server in the list. This increments the counter of error for the
|
* passed server, or creates an initial value of 1 error for it if the server
|
* is not yet present in the map.
|
* @param errorsByServer map of number of errors per serverID
|
* @param sid the ID of the server which produced an error
|
*/
|
private void updateAssuredErrorsByServer(Map<Integer,Integer> errorsByServer,
|
Integer sid)
|
{
|
synchronized (errorsByServer)
|
{
|
Integer serverErrCount = errorsByServer.get(sid);
|
if (serverErrCount == null)
|
{
|
// Server not present in list, create an entry with an
|
// initial number of errors set to 1
|
errorsByServer.put(sid, 1);
|
} else
|
{
|
// Server already present in list, just increment number of
|
// errors for the server
|
int val = serverErrCount;
|
val++;
|
errorsByServer.put(sid, val);
|
}
|
}
|
}
|
|
/**
|
* Do the necessary processing when an AckMsg is received.
|
*
|
* @param ack The AckMsg that was received.
|
*/
|
private void receiveAck(AckMsg ack)
|
{
|
CSN csn = ack.getCSN();
|
|
// Remove the message for pending ack list (this may already make the thread
|
// that is waiting for the ack be aware of its reception)
|
UpdateMsg update = waitingAckMsgs.remove(csn);
|
|
// Signal waiting thread ack has been received
|
if (update != null)
|
{
|
synchronized (update)
|
{
|
update.notify();
|
}
|
|
// Analyze status of embedded in the ack to see if everything went well
|
boolean hasTimeout = ack.hasTimeout();
|
boolean hasReplayErrors = ack.hasReplayError();
|
boolean hasWrongStatus = ack.hasWrongStatus();
|
|
AssuredMode updateAssuredMode = update.getAssuredMode();
|
|
if ( hasTimeout || hasReplayErrors || hasWrongStatus)
|
{
|
/*
|
Some problems detected: message did not correctly reach every
|
requested servers. Log problem
|
*/
|
Message errorMsg = NOTE_DS_RECEIVED_ACK_ERROR.get(
|
getBaseDNString(), Integer.toString(serverID),
|
update.toString(), ack.errorsToString());
|
logError(errorMsg);
|
|
List<Integer> failedServers = ack.getFailedServers();
|
|
// Increment assured replication monitoring counters
|
switch (updateAssuredMode)
|
{
|
case SAFE_READ_MODE:
|
assuredSrNotAcknowledgedUpdates.incrementAndGet();
|
if (hasTimeout)
|
assuredSrTimeoutUpdates.incrementAndGet();
|
if (hasReplayErrors)
|
assuredSrReplayErrorUpdates.incrementAndGet();
|
if (hasWrongStatus)
|
assuredSrWrongStatusUpdates.incrementAndGet();
|
if (failedServers != null) // This should always be the case !
|
{
|
for(Integer sid : failedServers)
|
{
|
updateAssuredErrorsByServer(
|
assuredSrServerNotAcknowledgedUpdates, sid);
|
}
|
}
|
break;
|
case SAFE_DATA_MODE:
|
// The only possible cause of ack error in safe data mode is timeout
|
if (hasTimeout) // So should always be the case
|
assuredSdTimeoutUpdates.incrementAndGet();
|
if (failedServers != null) // This should always be the case !
|
{
|
for(Integer sid : failedServers)
|
{
|
updateAssuredErrorsByServer(
|
assuredSdServerTimeoutUpdates, sid);
|
}
|
}
|
break;
|
default:
|
// Should not happen
|
}
|
} else
|
{
|
// Update has been acknowledged without errors
|
// Increment assured replication monitoring counters
|
switch (updateAssuredMode)
|
{
|
case SAFE_READ_MODE:
|
assuredSrAcknowledgedUpdates.incrementAndGet();
|
break;
|
case SAFE_DATA_MODE:
|
assuredSdAcknowledgedUpdates.incrementAndGet();
|
break;
|
default:
|
// Should not happen
|
}
|
}
|
}
|
}
|
|
|
/*
|
* After this point the code is related to the Total Update.
|
*/
|
|
/**
|
* This thread is launched when we want to export data to another server.
|
*
|
* When a task is created locally (so this local server is the initiator)
|
* of the export (Example: dsreplication initialize-all),
|
* this thread is NOT used but the task thread is running the export instead).
|
*/
|
private class ExportThread extends DirectoryThread
|
{
|
/** Id of server that will be initialized. */
|
private final int serverIdToInitialize;
|
private final int initWindow;
|
|
|
|
/**
|
* Constructor for the ExportThread.
|
*
|
* @param serverIdToInitialize
|
* serverId of server that will receive entries
|
* @param initWindow
|
* The value of the initialization window for flow control between
|
* the importer and the exporter.
|
*/
|
public ExportThread(int serverIdToInitialize, int initWindow)
|
{
|
super("Export thread from serverId=" + serverID + " to serverId="
|
+ serverIdToInitialize);
|
this.serverIdToInitialize = serverIdToInitialize;
|
this.initWindow = initWindow;
|
}
|
|
|
|
/**
|
* Run method for this class.
|
*/
|
@Override
|
public void run()
|
{
|
if (debugEnabled())
|
TRACER.debugInfo("[IE] starting " + this.getName());
|
try
|
{
|
initializeRemote(serverIdToInitialize, serverIdToInitialize, null,
|
initWindow);
|
} catch (DirectoryException de)
|
{
|
/*
|
An error message has been sent to the peer
|
This server is not the initiator of the export so there is
|
nothing more to do locally.
|
*/
|
}
|
|
if (debugEnabled())
|
TRACER.debugInfo("[IE] ending " + this.getName());
|
}
|
}
|
|
/**
|
* This class contain the context related to an import or export
|
* launched on the domain.
|
*/
|
protected class IEContext
|
{
|
/** The private task that initiated the operation. */
|
private Task initializeTask;
|
/** The destination in the case of an export. */
|
private int exportTarget = RoutableMsg.UNKNOWN_SERVER;
|
/** The source in the case of an import. */
|
private int importSource = RoutableMsg.UNKNOWN_SERVER;
|
|
/** The total entry count expected to be processed. */
|
private long entryCount = 0;
|
/** The count for the entry not yet processed. */
|
private long entryLeftCount = 0;
|
|
/** Exception raised during the initialization. */
|
private DirectoryException exception = null;
|
|
/** Whether the context is related to an import or an export. */
|
private boolean importInProgress;
|
|
/** Current counter of messages exchanged during the initialization. */
|
private int msgCnt = 0;
|
|
/**
|
* Number of connections lost when we start the initialization. Will help
|
* counting connections lost during initialization,
|
*/
|
private int initNumLostConnections = 0;
|
|
/**
|
* Request message sent when this server has the initializeFromRemote task.
|
*/
|
private InitializeRequestMsg initReqMsgSent = null;
|
|
/**
|
* Start time of the initialization process. ErrorMsg timestamped before
|
* this startTime will be ignored.
|
*/
|
private long startTime;
|
|
/**
|
* List for replicas (DS) connected to the topology when initialization
|
* started.
|
*/
|
private Set<Integer> startList = new HashSet<Integer>(0);
|
|
/**
|
* List for replicas (DS) with a failure (disconnected from the topology)
|
* since the initialization started.
|
*/
|
private Set<Integer> failureList = new HashSet<Integer>(0);
|
|
/**
|
* Flow control during initialization: for each remote server, counter of
|
* messages received.
|
*/
|
private final HashMap<Integer, Integer> ackVals =
|
new HashMap<Integer, Integer>();
|
/**
|
* ServerId of the slowest server (the one with the smallest non null
|
* counter).
|
*/
|
private int slowestServerId = -1;
|
|
private short exporterProtocolVersion = -1;
|
|
/** Window used during this initialization. */
|
private int initWindow;
|
|
/** Number of attempt already done for this initialization. */
|
private short attemptCnt;
|
|
/**
|
* Creates a new IEContext.
|
*
|
* @param importInProgress true if the IEContext will be used
|
* for and import, false if the IEContext
|
* will be used for and export.
|
*/
|
public IEContext(boolean importInProgress)
|
{
|
this.importInProgress = importInProgress;
|
this.startTime = System.currentTimeMillis();
|
this.attemptCnt = 0;
|
}
|
|
/**
|
* Initializes the import/export counters with the provider value.
|
* @param total Total number of entries to be processed.
|
* @throws DirectoryException if an error occurred.
|
*/
|
private void initializeCounters(long total)
|
throws DirectoryException
|
{
|
entryCount = total;
|
entryLeftCount = total;
|
|
if (initializeTask != null)
|
{
|
if (initializeTask instanceof InitializeTask)
|
{
|
((InitializeTask)initializeTask).setTotal(entryCount);
|
((InitializeTask)initializeTask).setLeft(entryCount);
|
}
|
else if (initializeTask instanceof InitializeTargetTask)
|
{
|
((InitializeTargetTask)initializeTask).setTotal(entryCount);
|
((InitializeTargetTask)initializeTask).setLeft(entryCount);
|
}
|
}
|
}
|
|
/**
|
* Update the counters of the task for each entry processed during
|
* an import or export.
|
*
|
* @param entriesDone The number of entries that were processed
|
* since the last time this method was called.
|
*
|
* @throws DirectoryException if an error occurred.
|
*/
|
public void updateCounters(int entriesDone)
|
throws DirectoryException
|
{
|
entryLeftCount -= entriesDone;
|
|
if (initializeTask != null)
|
{
|
if (initializeTask instanceof InitializeTask)
|
{
|
((InitializeTask)initializeTask).setLeft(entryLeftCount);
|
}
|
else if (initializeTask instanceof InitializeTargetTask)
|
{
|
((InitializeTargetTask)initializeTask).setLeft(entryLeftCount);
|
}
|
}
|
}
|
|
/**
|
* {@inheritDoc}
|
*/
|
@Override
|
public String toString()
|
{
|
return "[ Entry count=" + this.entryCount +
|
", Entry left count=" + this.entryLeftCount + "]";
|
}
|
|
/**
|
* Gets the server id of the exporting server.
|
* @return the server id of the exporting server.
|
*/
|
public int getExportTarget()
|
{
|
return exportTarget;
|
}
|
|
/**
|
* Gets the server id of the importing server.
|
* @return the server id of the importing server.
|
*/
|
public int getImportSource()
|
{
|
return importSource;
|
}
|
|
/**
|
* Get the exception that occurred during the import/export.
|
* @return the exception that occurred during the import/export.
|
*/
|
public DirectoryException getException()
|
{
|
return exception;
|
}
|
|
/**
|
* Set the exception that occurred during the import/export.
|
* @param exception the exception that occurred during the import/export.
|
*/
|
public void setException(DirectoryException exception)
|
{
|
this.exception = exception;
|
}
|
|
/**
|
* Set the id of the EntryMsg acknowledged from a receiver (importer)server.
|
* (updated via the listener thread)
|
* @param serverId serverId of the acknowledger/receiver/importer server.
|
* @param numAck id of the message received.
|
*/
|
public void setAckVal(int serverId, int numAck)
|
{
|
if (debugEnabled())
|
TRACER.debugInfo("[IE] setAckVal[" + serverId + "]=" + numAck);
|
|
this.ackVals.put(serverId, numAck);
|
|
// Recompute the server with the minAck returned,means the slowest server.
|
slowestServerId = serverId;
|
for (Integer sid : ieContext.ackVals.keySet())
|
if (this.ackVals.get(sid) < this.ackVals.get(slowestServerId))
|
slowestServerId = sid;
|
}
|
|
/**
|
* Returns the serverId of the server that acknowledged the smallest
|
* EntryMsg id.
|
* @return serverId of the server with latest acknowledge.
|
* 0 when no ack has been received yet.
|
*/
|
public int getSlowestServer()
|
{
|
if (debugEnabled())
|
TRACER.debugInfo("[IE] getSlowestServer" + slowestServerId
|
+ " " + this.ackVals.get(slowestServerId));
|
|
return this.slowestServerId;
|
}
|
}
|
/**
|
* Verifies that the given string represents a valid source
|
* from which this server can be initialized.
|
*
|
* @param targetString The string representing the source
|
* @return The source as a integer value
|
* @throws DirectoryException if the string is not valid
|
*/
|
public int decodeTarget(String targetString) throws DirectoryException
|
{
|
if (targetString.equalsIgnoreCase("all"))
|
{
|
return RoutableMsg.ALL_SERVERS;
|
}
|
|
// So should be a serverID
|
try
|
{
|
int target = Integer.decode(targetString);
|
if (target >= 0)
|
{
|
// FIXME Could we check now that it is a know server in the domain ?
|
}
|
return target;
|
}
|
catch (Exception e)
|
{
|
ResultCode resultCode = ResultCode.OTHER;
|
Message message = ERR_INVALID_EXPORT_TARGET.get();
|
throw new DirectoryException(resultCode, message, e);
|
}
|
}
|
|
/**
|
* Initializes a remote server from this server.
|
* <p>
|
* The {@code exportBackend(OutputStream)} will therefore be called
|
* on this server, and the {@code importBackend(InputStream)}
|
* will be called on the remote server.
|
* <p>
|
* The InputStream and OutputStream given as a parameter to those
|
* methods will be connected through the replication protocol.
|
*
|
* @param target The server-id of the server that should be initialized.
|
* The target can be discovered using the
|
* {@link #getReplicasList()} method.
|
* @param initTask The task that triggers this initialization and that should
|
* be updated with its progress.
|
*
|
* @throws DirectoryException If it was not possible to publish the
|
* Initialization message to the Topology.
|
*/
|
public void initializeRemote(int target, Task initTask)
|
throws DirectoryException
|
{
|
initializeRemote(target, this.serverID, initTask, this.initWindow);
|
}
|
|
/**
|
* Process the initialization of some other server or servers in the topology
|
* specified by the target argument when this initialization specifying the
|
* server that requests the initialization.
|
*
|
* @param serverToInitialize The target server that should be initialized.
|
* @param serverRunningTheTask The server that initiated the export. It can
|
* be the serverID of this server, or the serverID of a remote server.
|
* @param initTask The task in this server that triggers this initialization
|
* and that should be updated with its progress. Null when the export is done
|
* following a request coming from a remote server (task is remote).
|
* @param initWindow The value of the initialization window for flow control
|
* between the importer and the exporter.
|
*
|
* @exception DirectoryException When an error occurs. No exception raised
|
* means success.
|
*/
|
protected void initializeRemote(int serverToInitialize,
|
int serverRunningTheTask, Task initTask, int initWindow)
|
throws DirectoryException
|
{
|
DirectoryException exportRootException = null;
|
|
// Acquire and initialize the export context
|
acquireIEContext(false);
|
|
/*
|
We manage the list of servers to initialize in order :
|
- to test at the end that all expected servers have reconnected
|
after their import and with the right genId
|
- to update the task with the server(s) where this test failed
|
*/
|
|
if (serverToInitialize == RoutableMsg.ALL_SERVERS)
|
{
|
Message msg = NOTE_FULL_UPDATE_ENGAGED_FOR_REMOTE_START_ALL.get(
|
countEntries(), getBaseDNString(), serverID);
|
logError(msg);
|
|
for (DSInfo dsi : getReplicasList())
|
ieContext.startList.add(dsi.getDsId());
|
|
// We manage the list of servers with which a flow control can be enabled
|
for (DSInfo dsi : getReplicasList())
|
{
|
if (dsi.getProtocolVersion()>= ProtocolVersion.REPLICATION_PROTOCOL_V4)
|
ieContext.setAckVal(dsi.getDsId(), 0);
|
}
|
}
|
else
|
{
|
Message msg = NOTE_FULL_UPDATE_ENGAGED_FOR_REMOTE_START.get(
|
countEntries(), getBaseDNString(), serverID, serverToInitialize);
|
logError(msg);
|
|
ieContext.startList.add(serverToInitialize);
|
|
// We manage the list of servers with which a flow control can be enabled
|
for (DSInfo dsi : getReplicasList())
|
{
|
if (dsi.getDsId() == serverToInitialize &&
|
dsi.getProtocolVersion()>= ProtocolVersion.REPLICATION_PROTOCOL_V4)
|
ieContext.setAckVal(dsi.getDsId(), 0);
|
}
|
}
|
|
// loop for the case where the exporter is the initiator
|
int attempt = 0;
|
boolean done = false;
|
while (!done && ++attempt < 2) // attempt loop
|
{
|
try
|
{
|
ieContext.exportTarget = serverToInitialize;
|
if (initTask != null)
|
ieContext.initializeTask = initTask;
|
ieContext.initializeCounters(this.countEntries());
|
ieContext.msgCnt = 0;
|
ieContext.initNumLostConnections = broker.getNumLostConnections();
|
ieContext.initWindow = initWindow;
|
|
// Send start message to the peer
|
InitializeTargetMsg initTargetMsg = new InitializeTargetMsg(
|
getBaseDN(), serverID, serverToInitialize,
|
serverRunningTheTask, ieContext.entryCount, initWindow);
|
|
broker.publish(initTargetMsg);
|
|
// Wait for all servers to be ok
|
waitForRemoteStartOfInit();
|
|
// Servers that left in the list are those for which we could not test
|
// that they have been successfully initialized.
|
if (!ieContext.failureList.isEmpty())
|
{
|
throw new DirectoryException(
|
ResultCode.OTHER,
|
ERR_INIT_NO_SUCCESS_START_FROM_SERVERS.get(
|
ieContext.failureList.toString()));
|
}
|
|
exportBackend(new BufferedOutputStream(new ReplOutputStream(this)));
|
|
// Notify the peer of the success
|
DoneMsg doneMsg = new DoneMsg(serverID, initTargetMsg.getDestination());
|
broker.publish(doneMsg);
|
}
|
catch(DirectoryException exportException)
|
{
|
// Give priority to the first exception raised - stored in the context
|
if (ieContext.exception != null)
|
exportRootException = ieContext.exception;
|
else
|
exportRootException = exportException;
|
}
|
|
if (debugEnabled())
|
TRACER.debugInfo("[IE] In " + broker.getReplicationMonitorInstanceName()
|
+ " export ends with " + " connected=" + broker.isConnected()
|
+ " exportRootException=" + exportRootException);
|
|
if (exportRootException != null)
|
{
|
try
|
{
|
/*
|
Handling the errors during export
|
|
Note: we could have lost the connection and another thread
|
the listener one) has already managed to reconnect.
|
So we MUST rely on the test broker.isConnected()
|
ONLY to do 'wait to be reconnected by another thread'
|
(if not yet reconnected already).
|
*/
|
if (!broker.isConnected())
|
{
|
// We are still disconnected, so we wait for the listener thread
|
// to reconnect - wait 10s
|
if (debugEnabled())
|
TRACER.debugInfo(
|
"[IE] Exporter wait for reconnection by the listener thread");
|
int att=0;
|
while (!broker.shuttingDown() && !broker.isConnected()
|
&& ++att < 100)
|
try { Thread.sleep(100); }
|
catch(Exception e){ /* do nothing */ }
|
}
|
|
if (initTask != null && broker.isConnected()
|
&& serverToInitialize != RoutableMsg.ALL_SERVERS)
|
{
|
/*
|
NewAttempt case : In the case where
|
- it's not an InitializeAll
|
- AND the previous export attempt failed
|
- AND we are (now) connected
|
- and we own the task and this task is not an InitializeAll
|
Let's :
|
- sleep to let time to the other peer to reconnect if needed
|
- and launch another attempt
|
*/
|
try { Thread.sleep(1000); }
|
catch(Exception e){ /* do nothing */ }
|
logError(NOTE_RESENDING_INIT_TARGET.get(
|
exportRootException.getLocalizedMessage()));
|
|
continue;
|
}
|
|
ErrorMsg errorMsg =
|
new ErrorMsg(serverToInitialize,
|
exportRootException.getMessageObject());
|
broker.publish(errorMsg);
|
}
|
catch(Exception e)
|
{
|
// Ignore the failure raised while proceeding the root failure
|
}
|
}
|
|
// We are always done for this export ...
|
// ... except in the NewAttempt case (see above)
|
done = true;
|
|
} // attempt loop
|
|
// Wait for all servers to be ok, and build the failure list
|
waitForRemoteEndOfInit();
|
|
// Servers that left in the list are those for which we could not test
|
// that they have been successfully initialized.
|
if (!ieContext.failureList.isEmpty() && exportRootException == null)
|
{
|
exportRootException = new DirectoryException(ResultCode.OTHER,
|
ERR_INIT_NO_SUCCESS_END_FROM_SERVERS.get(
|
Long.toString(getGenerationID()),
|
ieContext.failureList.toString()));
|
}
|
|
// Don't forget to release IEcontext acquired at beginning.
|
releaseIEContext();
|
|
String cause = exportRootException != null ? exportRootException
|
.getLocalizedMessage() : "";
|
if (serverToInitialize == RoutableMsg.ALL_SERVERS)
|
{
|
Message msg = NOTE_FULL_UPDATE_ENGAGED_FOR_REMOTE_END_ALL.get(
|
getBaseDNString(), serverID, cause);
|
logError(msg);
|
}
|
else
|
{
|
Message msg = NOTE_FULL_UPDATE_ENGAGED_FOR_REMOTE_END.get(
|
getBaseDNString(), serverID, serverToInitialize, cause);
|
logError(msg);
|
}
|
|
|
if (exportRootException != null)
|
{
|
throw exportRootException;
|
}
|
}
|
|
/**
|
* For all remote servers in the start list:
|
* - wait it has finished the import and present the expected generationID,
|
* - build the failureList.
|
*/
|
private void waitForRemoteStartOfInit()
|
{
|
int waitResultAttempt = 0;
|
Set<Integer> replicasWeAreWaitingFor = new HashSet<Integer>(0);
|
|
replicasWeAreWaitingFor.addAll(ieContext.startList);
|
|
if (debugEnabled())
|
TRACER.debugInfo(
|
"[IE] wait for start replicasWeAreWaitingFor=" + replicasWeAreWaitingFor);
|
|
boolean done;
|
do
|
{
|
done = true;
|
for (DSInfo dsi : getReplicasList())
|
{
|
if (debugEnabled())
|
TRACER.debugInfo(
|
"[IE] wait for start dsid " + dsi.getDsId()
|
+ " " + dsi.getStatus()
|
+ " " + dsi.getGenerationId()
|
+ " " + this.getGenerationID());
|
if (ieContext.startList.contains(dsi.getDsId()))
|
{
|
if (dsi.getStatus() != ServerStatus.FULL_UPDATE_STATUS)
|
{
|
// this one is still not doing the Full Update ... retry later
|
done = false;
|
try { Thread.sleep(100);
|
}
|
catch (InterruptedException e) {
|
Thread.currentThread().interrupt();
|
}
|
waitResultAttempt++;
|
break;
|
}
|
else
|
{
|
// this one is ok
|
replicasWeAreWaitingFor.remove(dsi.getDsId());
|
}
|
}
|
}
|
}
|
while (!done && waitResultAttempt < 1200 && !broker.shuttingDown());
|
|
ieContext.failureList.addAll(replicasWeAreWaitingFor);
|
|
if (debugEnabled())
|
TRACER.debugInfo(
|
"[IE] wait for start ends with " + ieContext.failureList);
|
}
|
|
/**
|
* For all remote servers in the start list:
|
* - wait it has finished the import and present the expected generationID,
|
* - build the failureList.
|
*/
|
private void waitForRemoteEndOfInit()
|
{
|
Set<Integer> replicasWeAreWaitingFor = new HashSet<Integer>(
|
ieContext.startList);
|
|
if (debugEnabled())
|
TRACER.debugInfo(
|
"[IE] wait for end replicasWeAreWaitingFor=" + replicasWeAreWaitingFor);
|
|
/*
|
In case some new servers appear during the init, we want them to be
|
considered in the processing of sorting the successfully initialized
|
and the others
|
*/
|
for (DSInfo dsi : getReplicasList())
|
replicasWeAreWaitingFor.add(dsi.getDsId());
|
|
boolean done;
|
do
|
{
|
done = true;
|
short reconnectMaxDelayInSec = 10;
|
short reconnectWait = 0;
|
Iterator<Integer> it = replicasWeAreWaitingFor.iterator();
|
while (it.hasNext())
|
{
|
int serverId = it.next();
|
if (ieContext.failureList.contains(serverId))
|
{
|
/*
|
this server has already been in error during initialization
|
don't wait for it
|
*/
|
continue;
|
}
|
|
DSInfo dsInfo = isRemoteDSConnected(serverId);
|
if (dsInfo == null)
|
{
|
/*
|
this server is disconnected
|
may be for a long time if it crashed or had been stopped
|
may be just the time to reconnect after import : should be short
|
*/
|
if (++reconnectWait<reconnectMaxDelayInSec)
|
{
|
// let's still wait to give a chance to this server to reconnect
|
done = false;
|
}
|
// Else we left enough time to the servers to reconnect
|
}
|
else
|
{
|
// this server is connected
|
if (dsInfo.getStatus() == ServerStatus.FULL_UPDATE_STATUS)
|
{
|
// this one is still doing the Full Update ... retry later
|
done = false;
|
break;
|
}
|
else
|
{
|
if (dsInfo.getGenerationId() == this.getGenerationID())
|
{ // and with the expected generationId
|
// We're done with this server
|
it.remove();
|
}
|
}
|
}
|
}
|
|
// loop and wait
|
if (!done)
|
try { Thread.sleep(1000); }
|
catch (InterruptedException e) {
|
Thread.currentThread().interrupt();
|
} // 1sec
|
|
}
|
while (!done && !broker.shuttingDown()); // infinite wait
|
|
ieContext.failureList.addAll(replicasWeAreWaitingFor);
|
|
if (debugEnabled())
|
TRACER.debugInfo(
|
"[IE] wait for end ends with " + ieContext.failureList);
|
|
|
}
|
|
/**
|
* Get the ServerState maintained by the Concrete class.
|
*
|
* @return the ServerState maintained by the Concrete class.
|
*/
|
public ServerState getServerState()
|
{
|
return state;
|
}
|
|
|
private synchronized void acquireIEContext(boolean importInProgress)
|
throws DirectoryException
|
{
|
if (ieContext != null)
|
{
|
// Rejects 2 simultaneous exports
|
Message message = ERR_SIMULTANEOUS_IMPORT_EXPORT_REJECTED.get();
|
throw new DirectoryException(ResultCode.OTHER,
|
message);
|
}
|
|
ieContext = new IEContext(importInProgress);
|
}
|
|
private synchronized void releaseIEContext()
|
{
|
ieContext = null;
|
}
|
|
/**
|
* Processes an error message received while an export is
|
* on going, or an import will start.
|
*
|
* @param errorMsg The error message received.
|
*/
|
private void processErrorMsg(ErrorMsg errorMsg)
|
{
|
if (ieContext != null)
|
{
|
/*
|
Exporting must not be stopped on the first error, if we
|
run initialize-all.
|
*/
|
if (ieContext.exportTarget != RoutableMsg.ALL_SERVERS)
|
{
|
// The ErrorMsg is received while we have started an initialization
|
if (ieContext.getException() == null)
|
ieContext.setException(new DirectoryException(ResultCode.OTHER,
|
errorMsg.getDetails()));
|
|
/*
|
* This can happen :
|
* - on the first InitReqMsg sent when source in not known for example
|
* - on the next attempt when source crashed and did not reconnect
|
* even after the nextInitAttemptDelay
|
* During the import, the ErrorMsg will be received by receiveEntryBytes
|
*/
|
if (ieContext.initializeTask instanceof InitializeTask)
|
{
|
// Update the task that initiated the import
|
((InitializeTask)ieContext.initializeTask).
|
updateTaskCompletionState(ieContext.getException());
|
|
releaseIEContext();
|
}
|
}
|
}
|
}
|
|
/**
|
* Receives bytes related to an entry in the context of an import to
|
* initialize the domain (called by ReplLDIFInputStream).
|
*
|
* @return The bytes. Null when the Done or Err message has been received
|
*/
|
protected byte[] receiveEntryBytes()
|
{
|
ReplicationMsg msg;
|
while (true)
|
{
|
try
|
{
|
// In the context of the total update, we don't want any automatic
|
// re-connection done transparently by the broker because of a better
|
// RS or because of a connection failure.
|
// We want to be notified of topology change in order to track a
|
// potential disconnection of the exporter.
|
msg = broker.receive(false, false, true);
|
|
if (debugEnabled())
|
TRACER.debugInfo("[IE] In "
|
+ broker.getReplicationMonitorInstanceName()
|
+ ", receiveEntryBytes " + msg);
|
|
if (msg == null)
|
{
|
if (broker.shuttingDown())
|
{
|
// The server is in the shutdown process
|
return null;
|
}
|
else
|
{
|
// Handle connection issues
|
if (ieContext.getException() == null)
|
ieContext.setException(new DirectoryException(
|
ResultCode.OTHER,
|
ERR_INIT_RS_DISCONNECTION_DURING_IMPORT.get(
|
broker.getReplicationServer())));
|
return null;
|
}
|
}
|
|
// Check good ordering of msg received
|
if (msg instanceof EntryMsg)
|
{
|
EntryMsg entryMsg = (EntryMsg)msg;
|
byte[] entryBytes = entryMsg.getEntryBytes();
|
ieContext.updateCounters(countEntryLimits(entryBytes));
|
|
if (ieContext.exporterProtocolVersion >=
|
ProtocolVersion.REPLICATION_PROTOCOL_V4)
|
{
|
// check the msgCnt of the msg received to check ordering
|
if (++ieContext.msgCnt != entryMsg.getMsgId())
|
{
|
if (ieContext.getException() == null)
|
ieContext.setException(new DirectoryException(ResultCode.OTHER,
|
ERR_INIT_BAD_MSG_ID_SEQ_DURING_IMPORT.get(
|
String.valueOf(ieContext.msgCnt),
|
String.valueOf(entryMsg.getMsgId()))));
|
return null;
|
}
|
|
// send the ack of flow control mgmt
|
if ((ieContext.msgCnt % (ieContext.initWindow/2)) == 0)
|
{
|
InitializeRcvAckMsg amsg = new InitializeRcvAckMsg(
|
this.serverID,
|
entryMsg.getSenderID(),
|
ieContext.msgCnt);
|
broker.publish(amsg, false);
|
if (debugEnabled())
|
TRACER.debugInfo("[IE] In "
|
+ broker.getReplicationMonitorInstanceName()
|
+ ", publish InitializeRcvAckMsg" + amsg);
|
}
|
}
|
return entryBytes;
|
}
|
else if (msg instanceof DoneMsg)
|
{
|
/*
|
This is the normal termination of the import
|
No error is stored and the import is ended
|
by returning null
|
*/
|
return null;
|
}
|
else if (msg instanceof ErrorMsg)
|
{
|
/*
|
This is an error termination during the import
|
The error is stored and the import is ended
|
by returning null
|
*/
|
if (ieContext.getException() == null)
|
{
|
ErrorMsg errMsg = (ErrorMsg)msg;
|
if (errMsg.getCreationTime() > ieContext.startTime)
|
{
|
ieContext.setException(
|
new DirectoryException(ResultCode.OTHER,errMsg.getDetails()));
|
return null;
|
}
|
}
|
}
|
else
|
{
|
// Other messages received during an import are trashed except
|
// the topologyMsg.
|
if ((msg instanceof TopologyMsg) &&
|
(isRemoteDSConnected(ieContext.importSource)==null))
|
{
|
Message errMsg =
|
Message.raw(Category.SYNC, Severity.NOTICE,
|
ERR_INIT_EXPORTER_DISCONNECTION.get(
|
getBaseDNString(),
|
Integer.toString(this.serverID),
|
Integer.toString(ieContext.importSource)));
|
if (ieContext.getException()==null)
|
ieContext.setException(new DirectoryException(ResultCode.OTHER,
|
errMsg));
|
return null;
|
}
|
}
|
}
|
catch(Exception e)
|
{
|
if (ieContext.getException() == null)
|
ieContext.setException(new DirectoryException(ResultCode.OTHER,
|
ERR_INIT_IMPORT_FAILURE.get(e.getLocalizedMessage())));
|
}
|
}
|
}
|
|
/**
|
* Count the number of entries in the provided byte[].
|
* This is based on the hypothesis that the entries are separated
|
* by a "\n\n" String.
|
*
|
* @param entryBytes the set of bytes containing one or more entries.
|
* @return The number of entries in the provided byte[].
|
*/
|
private int countEntryLimits(byte[] entryBytes)
|
{
|
return countEntryLimits(entryBytes, 0, entryBytes.length);
|
}
|
|
/**
|
* Count the number of entries in the provided byte[].
|
* This is based on the hypothesis that the entries are separated
|
* by a "\n\n" String.
|
*
|
* @param entryBytes the set of bytes containing one or more entries.
|
* @return The number of entries in the provided byte[].
|
*/
|
private int countEntryLimits(byte[] entryBytes, int pos, int length)
|
{
|
int entryCount = 0;
|
int count = 0;
|
while (count<=length-2)
|
{
|
if ((entryBytes[pos+count] == '\n') && (entryBytes[pos+count+1] == '\n'))
|
{
|
entryCount++;
|
count++;
|
}
|
count++;
|
}
|
return entryCount;
|
}
|
|
/**
|
* Exports an entry in LDIF format.
|
*
|
* @param lDIFEntry The entry to be exported in byte[] form.
|
* @param pos The starting Position in the array.
|
* @param length Number of array elements to be copied.
|
*
|
* @throws IOException when an error occurred.
|
*/
|
public void exportLDIFEntry(byte[] lDIFEntry, int pos, int length)
|
throws IOException
|
{
|
if (debugEnabled())
|
TRACER.debugInfo("[IE] Entering exportLDIFEntry entry=" +
|
Arrays.toString(lDIFEntry));
|
|
// build the message
|
EntryMsg entryMessage = new EntryMsg(
|
serverID,ieContext.getExportTarget(), lDIFEntry, pos, length,
|
++ieContext.msgCnt);
|
|
// Waiting the slowest loop
|
while (!broker.shuttingDown())
|
{
|
/*
|
If an error was raised - like receiving an ErrorMsg from a remote
|
server that have been stored by the listener thread in the ieContext,
|
we just abandon the export by throwing an exception.
|
*/
|
if (ieContext.getException() != null)
|
throw new IOException(ieContext.getException().getMessage());
|
|
int slowestServerId = ieContext.getSlowestServer();
|
if (isRemoteDSConnected(slowestServerId)==null)
|
{
|
ieContext.setException(new DirectoryException(ResultCode.OTHER,
|
ERR_INIT_HEARTBEAT_LOST_DURING_EXPORT.get(
|
Integer.toString(ieContext.getSlowestServer()))));
|
|
throw new IOException("IOException with nested DirectoryException",
|
ieContext.getException());
|
}
|
|
int ourLastExportedCnt = ieContext.msgCnt;
|
int slowestCnt = ieContext.ackVals.get(slowestServerId);
|
|
if (debugEnabled())
|
TRACER.debugInfo("[IE] Entering exportLDIFEntry waiting " +
|
" our=" + ourLastExportedCnt + " slowest=" + slowestCnt);
|
|
if ((ourLastExportedCnt - slowestCnt) > ieContext.initWindow)
|
{
|
if (debugEnabled())
|
TRACER.debugInfo("[IE] Entering exportLDIFEntry waiting");
|
|
// our export is too far beyond the slowest importer - let's wait
|
try { Thread.sleep(100); }
|
catch(Exception e) { /* do nothing */ }
|
|
// process any connection error
|
if ((broker.hasConnectionError())||
|
(broker.getNumLostConnections()!= ieContext.initNumLostConnections))
|
{
|
// publish failed - store the error in the ieContext ...
|
DirectoryException de = new DirectoryException(ResultCode.OTHER,
|
ERR_INIT_RS_DISCONNECTION_DURING_EXPORT.get(
|
Integer.toString(broker.getRsServerId())));
|
if (ieContext.getException() == null)
|
ieContext.setException(de);
|
// .. and abandon the export by throwing an exception.
|
throw new IOException(de.getMessage());
|
}
|
}
|
else
|
{
|
if (debugEnabled())
|
TRACER.debugInfo("[IE] slowest got to us => stop waiting");
|
break;
|
}
|
} // Waiting the slowest loop
|
|
if (debugEnabled())
|
TRACER.debugInfo("[IE] Entering exportLDIFEntry pub entry="
|
+ Arrays.toString(lDIFEntry));
|
|
boolean sent = broker.publish(entryMessage, false);
|
|
// process any publish error
|
if (!sent
|
|| broker.hasConnectionError()
|
|| broker.getNumLostConnections() != ieContext.initNumLostConnections)
|
{
|
// publish failed - store the error in the ieContext ...
|
DirectoryException de = new DirectoryException(ResultCode.OTHER,
|
ERR_INIT_RS_DISCONNECTION_DURING_EXPORT.get(
|
Integer.toString(broker.getRsServerId())));
|
if (ieContext.getException() == null)
|
ieContext.setException(de);
|
// .. and abandon the export by throwing an exception.
|
throw new IOException(de.getMessage());
|
}
|
|
// publish succeeded
|
try
|
{
|
ieContext.updateCounters(countEntryLimits(lDIFEntry, pos, length));
|
}
|
catch (DirectoryException de)
|
{
|
// store the error in the ieContext ...
|
if (ieContext.getException() == null)
|
ieContext.setException(de);
|
// .. and abandon the export by throwing an exception.
|
throw new IOException(de.getMessage());
|
}
|
}
|
|
/**
|
* Initializes this domain from another source server.
|
* <p>
|
* When this method is called, a request for initialization will
|
* be sent to the source server asking for initialization.
|
* <p>
|
* The {@code exportBackend(OutputStream)} will therefore be called
|
* on the source server, and the {@code importBackend(InputStream)}
|
* will be called on his server.
|
* <p>
|
* The InputStream and OutpuStream given as a parameter to those
|
* methods will be connected through the replication protocol.
|
*
|
* @param source The server-id of the source from which to initialize.
|
* The source can be discovered using the
|
* {@link #getReplicasList()} method.
|
*
|
* @throws DirectoryException If it was not possible to publish the
|
* Initialization message to the Topology.
|
*/
|
public void initializeFromRemote(int source) throws DirectoryException
|
{
|
initializeFromRemote(source, null);
|
}
|
|
/**
|
* Initializes a remote server from this server.
|
* <p>
|
* The {@code exportBackend(OutputStream)} will therefore be called
|
* on this server, and the {@code importBackend(InputStream)}
|
* will be called on the remote server.
|
* <p>
|
* The InputStream and OutpuStream given as a parameter to those
|
* methods will be connected through the replication protocol.
|
*
|
* @param target The server-id of the server that should be initialized.
|
* The target can be discovered using the
|
* {@link #getReplicasList()} method.
|
*
|
* @throws DirectoryException If it was not possible to publish the
|
* Initialization message to the Topology.
|
*/
|
public void initializeRemote(int target) throws DirectoryException
|
{
|
initializeRemote(target, null);
|
}
|
|
/**
|
* Initializes asynchronously this domain from a remote source server.
|
* Before returning from this call, for the provided task :
|
* - the progressing counters are updated during the initialization using
|
* setTotal() and setLeft().
|
* - the end of the initialization using updateTaskCompletionState().
|
* <p>
|
* When this method is called, a request for initialization is sent to the
|
* remote source server requesting initialization.
|
* <p>
|
*
|
* @param source The server-id of the source from which to initialize.
|
* The source can be discovered using the
|
* {@link #getReplicasList()} method.
|
*
|
* @param initTask The task that launched the initialization
|
* and should be updated of its progress.
|
*
|
* @throws DirectoryException If it was not possible to publish the
|
* Initialization message to the Topology.
|
* The task state is updated.
|
*/
|
public void initializeFromRemote(int source, Task initTask)
|
throws DirectoryException
|
{
|
Message errMsg = null;
|
|
if (debugEnabled())
|
TRACER.debugInfo("[IE] Entering initializeFromRemote for " + this);
|
|
if (!broker.isConnected())
|
{
|
errMsg = ERR_INITIALIZATION_FAILED_NOCONN.get(getBaseDNString());
|
}
|
|
/*
|
We must not test here whether the remote source is connected to
|
the topology by testing if it stands in the replicas list since.
|
In the case of a re-attempt of initialization, the listener thread is
|
running this method directly coming from initialize() method and did
|
not processed any topology message in between the failure and the
|
new attempt.
|
*/
|
try
|
{
|
/*
|
We must immediately acquire a context to store the task inside
|
The context will be used when we (the listener thread) will receive
|
the InitializeTargetMsg, process the import, and at the end
|
update the task.
|
*/
|
|
acquireIEContext(true); //test and set if no import already in progress
|
ieContext.initializeTask = initTask;
|
ieContext.attemptCnt = 0;
|
ieContext.initReqMsgSent = new InitializeRequestMsg(
|
getBaseDN(), serverID, source, this.initWindow);
|
|
// Publish Init request msg
|
broker.publish(ieContext.initReqMsgSent);
|
|
/*
|
The normal success processing is now to receive InitTargetMsg then
|
entries from the remote server.
|
The error cases are :
|
- either local error immediately caught below
|
- a remote error we will receive as an ErrorMsg
|
*/
|
}
|
catch(DirectoryException de)
|
{
|
errMsg = de.getMessageObject();
|
}
|
catch(Exception e)
|
{
|
// Should not happen
|
errMsg = Message.raw(Category.SYNC, Severity.NOTICE,
|
e.getLocalizedMessage());
|
logError(errMsg);
|
}
|
|
// When error, update the task and raise the error to the caller
|
if (errMsg != null)
|
{
|
// No need to call here updateTaskCompletionState - will be done
|
// by the caller
|
releaseIEContext();
|
throw new DirectoryException(ResultCode.OTHER, errMsg);
|
}
|
}
|
|
/**
|
* Processes an InitializeTargetMsg received from a remote server
|
* meaning processes an initialization from the entries expected to be
|
* received now.
|
*
|
* @param initTargetMsgReceived The message received from the remote server.
|
*
|
* @param requesterServerId The serverId of the server that requested the
|
* initialization meaning the server where the
|
* task has initially been created (this server,
|
* or the remote server).
|
*/
|
void initialize(InitializeTargetMsg initTargetMsgReceived,
|
int requesterServerId)
|
{
|
InitializeTask initFromTask = null;
|
|
if (debugEnabled())
|
TRACER.debugInfo("[IE] Entering initialize - domain=" + this);
|
|
int source = initTargetMsgReceived.getSenderID();
|
|
try
|
{
|
// Log starting
|
Message msg = NOTE_FULL_UPDATE_ENGAGED_FROM_REMOTE_START.get(
|
getBaseDNString(), initTargetMsgReceived.getSenderID(), serverID);
|
logError(msg);
|
|
// Go into full update status
|
setNewStatus(StatusMachineEvent.TO_FULL_UPDATE_STATUS_EVENT);
|
|
// Acquire an import context if no already done (and initialize).
|
if (initTargetMsgReceived.getInitiatorID() != this.serverID)
|
{
|
/*
|
The initTargetMsgReceived is for an import initiated by the remote
|
server.
|
Test and set if no import already in progress
|
*/
|
acquireIEContext(true);
|
}
|
|
// Initialize stuff
|
ieContext.importSource = source;
|
ieContext.initializeCounters(initTargetMsgReceived.getEntryCount());
|
ieContext.initWindow = initTargetMsgReceived.getInitWindow();
|
// Protocol version is -1 when not known.
|
ieContext.exporterProtocolVersion = getProtocolVersion(source);
|
initFromTask = (InitializeTask)ieContext.initializeTask;
|
|
// Launch the import
|
importBackend(new ReplInputStream(this));
|
|
}
|
catch (DirectoryException e)
|
{
|
/*
|
Store the exception raised. It will be considered if no other exception
|
has been previously stored in the context
|
*/
|
if (ieContext.getException() == null)
|
ieContext.setException(e);
|
}
|
finally
|
{
|
if (debugEnabled())
|
TRACER.debugInfo("[IE] Domain=" + this
|
+ " ends import with exception=" + ieContext.getException()
|
+ " connected=" + broker.isConnected());
|
|
/*
|
It is necessary to restart (reconnect to RS) for different reasons
|
- when everything went well, reconnect in order to exchange
|
new state, new generation ID
|
- when we have connection failure, reconnect to retry a new import
|
right here, right now
|
we never want retryOnFailure if we fails reconnecting in the restart.
|
*/
|
broker.reStart(false);
|
|
if (ieContext.getException() != null
|
&& broker.isConnected()
|
&& initFromTask != null
|
&& ++ieContext.attemptCnt < 2)
|
{
|
/*
|
Worth a new attempt
|
since initFromTask is in this server, connection is ok
|
*/
|
try
|
{
|
/*
|
Wait for the exporter to stabilize - eventually reconnect as
|
well if it was connected to the same RS than the one we lost ...
|
*/
|
Thread.sleep(1000);
|
|
/*
|
Restart the whole import protocol exchange by sending again
|
the request
|
*/
|
logError(NOTE_RESENDING_INIT_FROM_REMOTE_REQUEST.get(
|
ieContext.getException().getLocalizedMessage()));
|
|
broker.publish(ieContext.initReqMsgSent);
|
|
ieContext.initializeCounters(0);
|
ieContext.exception = null;
|
ieContext.msgCnt = 0;
|
|
// Processing of the received initTargetMsgReceived is done
|
// let's wait for the next one
|
return;
|
}
|
catch(Exception e)
|
{
|
/*
|
An error occurs when sending a new request for a new import.
|
This error is not stored, preferring to keep the initial one.
|
*/
|
logError(ERR_SENDING_NEW_ATTEMPT_INIT_REQUEST.get(
|
e.getLocalizedMessage(),
|
ieContext.getException().getLocalizedMessage()));
|
}
|
}
|
|
// ===================
|
// No new attempt case
|
|
if (debugEnabled())
|
TRACER.debugInfo("[IE] Domain=" + this
|
+ " ends initialization with exception=" + ieContext.getException()
|
+ " connected=" + broker.isConnected()
|
+ " task=" + initFromTask
|
+ " attempt=" + ieContext.attemptCnt);
|
|
try
|
{
|
if (broker.isConnected() && ieContext.getException() != null)
|
{
|
// Let's notify the exporter
|
ErrorMsg errorMsg = new ErrorMsg(requesterServerId,
|
ieContext.getException().getMessageObject());
|
broker.publish(errorMsg);
|
}
|
/*
|
Update the task that initiated the import must be the last thing.
|
Particularly, broker.restart() after import success must be done
|
before some other operations/tasks to be launched,
|
like resetting the generation ID.
|
*/
|
if (initFromTask != null)
|
{
|
initFromTask.updateTaskCompletionState(ieContext.getException());
|
}
|
}
|
finally
|
{
|
Message msg = NOTE_FULL_UPDATE_ENGAGED_FROM_REMOTE_END.get(
|
getBaseDNString(), initTargetMsgReceived.getSenderID(), serverID,
|
(ieContext.getException() != null ? ieContext
|
.getException().getLocalizedMessage() : ""));
|
logError(msg);
|
releaseIEContext();
|
} // finally
|
} // finally
|
}
|
|
/**
|
* Return the protocol version of the DS related to the provided serverId.
|
* Returns -1 when the protocol version is not known.
|
* @param dsServerId The provided serverId.
|
* @return The protocol version.
|
*/
|
short getProtocolVersion(int dsServerId)
|
{
|
short protocolVersion = -1;
|
for (DSInfo dsi : getReplicasList())
|
{
|
if (dsi.getDsId() == dsServerId)
|
{
|
protocolVersion = dsi.getProtocolVersion();
|
break;
|
}
|
}
|
return protocolVersion;
|
}
|
|
/**
|
* Sets the status to a new value depending of the passed status machine
|
* event.
|
* @param event The event that may make the status be changed
|
*/
|
protected void setNewStatus(StatusMachineEvent event)
|
{
|
ServerStatus newStatus =
|
StatusMachine.computeNewStatus(status, event);
|
|
if (newStatus == ServerStatus.INVALID_STATUS)
|
{
|
Message msg = ERR_DS_CANNOT_CHANGE_STATUS.get(getBaseDNString(),
|
Integer.toString(serverID), status.toString(), event.toString());
|
logError(msg);
|
return;
|
}
|
|
if (newStatus != status)
|
{
|
// Reset status date
|
lastStatusChangeDate = new Date();
|
// Reset monitoring counters if reconnection
|
if (newStatus == ServerStatus.NOT_CONNECTED_STATUS)
|
resetMonitoringCounters();
|
|
// Store new status
|
status = newStatus;
|
|
if (debugEnabled())
|
TRACER.debugInfo("Replication domain " + baseDN + " new status is: "
|
+ status);
|
|
// Perform whatever actions are needed to apply properties for being
|
// compliant with new status
|
updateDomainForNewStatus();
|
}
|
}
|
|
/**
|
* Returns a boolean indicating if an import or export is currently
|
* processed.
|
*
|
* @return The status
|
*/
|
public boolean ieRunning()
|
{
|
return ieContext != null;
|
}
|
|
/**
|
* Check the value of the Replication Servers generation ID.
|
*
|
* @param generationID The expected value of the generation ID.
|
*
|
* @throws DirectoryException When the generation ID of the Replication
|
* Servers is not the expected value.
|
*/
|
private void checkGenerationID(long generationID)
|
throws DirectoryException
|
{
|
boolean allSet = true;
|
|
for (int i = 0; i< 50; i++)
|
{
|
allSet = true;
|
for (RSInfo rsInfo : getRsList())
|
{
|
// the 'empty' RSes (generationId==-1) are considered as good citizens
|
if (rsInfo.getGenerationId() != -1 &&
|
rsInfo.getGenerationId() != generationID)
|
{
|
try
|
{
|
Thread.sleep(i*100);
|
} catch (InterruptedException e)
|
{
|
Thread.currentThread().interrupt();
|
}
|
allSet = false;
|
break;
|
}
|
}
|
if (allSet)
|
{
|
break;
|
}
|
}
|
if (!allSet)
|
{
|
Message message = ERR_RESET_GENERATION_ID_FAILED.get(getBaseDNString());
|
throw new DirectoryException(ResultCode.OTHER, message);
|
}
|
}
|
|
/**
|
* Reset the Replication Log.
|
* Calling this method will remove all the Replication information that
|
* was kept on all the Replication Servers currently connected in the
|
* topology.
|
*
|
* @throws DirectoryException If this ReplicationDomain is not currently
|
* connected to a Replication Server or it
|
* was not possible to contact it.
|
*/
|
public void resetReplicationLog() throws DirectoryException
|
{
|
// Reset the Generation ID to -1 to clean the ReplicationServers.
|
resetGenerationId((long)-1);
|
|
// check that at least one ReplicationServer did change its generation-id
|
checkGenerationID(-1);
|
|
// Reconnect to the Replication Server so that it adopt our
|
// GenerationID.
|
disableService();
|
enableService();
|
|
// wait for the domain to reconnect.
|
int count = 0;
|
while (!isConnected() && count < 10)
|
{
|
try
|
{
|
Thread.sleep(100);
|
} catch (InterruptedException e)
|
{
|
Thread.currentThread().interrupt();
|
}
|
}
|
|
resetGenerationId(getGenerationID());
|
|
// check that at least one ReplicationServer did change its generation-id
|
checkGenerationID(getGenerationID());
|
}
|
|
/**
|
* Reset the generationId of this domain in the whole topology.
|
* A message is sent to the Replication Servers for them to reset
|
* their change dbs.
|
*
|
* @param generationIdNewValue The new value of the generation Id.
|
* @throws DirectoryException When an error occurs
|
*/
|
public void resetGenerationId(Long generationIdNewValue)
|
throws DirectoryException
|
{
|
if (debugEnabled())
|
TRACER.debugInfo("Server id " + serverID + " and domain " + baseDN
|
+ " resetGenerationId " + generationIdNewValue);
|
|
ResetGenerationIdMsg genIdMessage;
|
|
if (generationIdNewValue == null)
|
{
|
genIdMessage = new ResetGenerationIdMsg(this.getGenerationID());
|
}
|
else
|
{
|
genIdMessage = new ResetGenerationIdMsg(generationIdNewValue);
|
}
|
|
if (!isConnected())
|
{
|
ResultCode resultCode = ResultCode.OTHER;
|
Message message = ERR_RESET_GENERATION_CONN_ERR_ID.get(getBaseDNString(),
|
Integer.toString(serverID),
|
Long.toString(genIdMessage.getGenerationId()));
|
throw new DirectoryException(
|
resultCode, message);
|
}
|
broker.publish(genIdMessage);
|
|
// check that at least one ReplicationServer did change its generation-id
|
if (generationIdNewValue == null)
|
{
|
checkGenerationID(this.getGenerationID());
|
}
|
else
|
{
|
checkGenerationID(generationIdNewValue);
|
}
|
}
|
|
|
/*
|
******** End of The total Update code *********
|
*/
|
|
/*
|
******* Start of Monitoring Code **********
|
*/
|
|
/**
|
* Get the maximum receive window size.
|
*
|
* @return The maximum receive window size.
|
*/
|
int getMaxRcvWindow()
|
{
|
if (broker != null)
|
return broker.getMaxRcvWindow();
|
return 0;
|
}
|
|
/**
|
* Get the current receive window size.
|
*
|
* @return The current receive window size.
|
*/
|
int getCurrentRcvWindow()
|
{
|
if (broker != null)
|
return broker.getCurrentRcvWindow();
|
return 0;
|
}
|
|
/**
|
* Get the maximum send window size.
|
*
|
* @return The maximum send window size.
|
*/
|
int getMaxSendWindow()
|
{
|
if (broker != null)
|
return broker.getMaxSendWindow();
|
return 0;
|
}
|
|
/**
|
* Get the current send window size.
|
*
|
* @return The current send window size.
|
*/
|
int getCurrentSendWindow()
|
{
|
if (broker != null)
|
return broker.getCurrentSendWindow();
|
return 0;
|
}
|
|
/**
|
* Get the number of times the replication connection was lost.
|
* @return The number of times the replication connection was lost.
|
*/
|
int getNumLostConnections()
|
{
|
if (broker != null)
|
return broker.getNumLostConnections();
|
return 0;
|
}
|
|
/**
|
* Determine whether the connection to the replication server is encrypted.
|
* @return true if the connection is encrypted, false otherwise.
|
*/
|
boolean isSessionEncrypted()
|
{
|
return broker != null && broker.isSessionEncrypted();
|
}
|
|
/**
|
* This method is called when the ReplicationDomain has completed the
|
* processing of a received update synchronously.
|
* In such cases the processUpdateDone () is called and the state
|
* is updated automatically.
|
*
|
* @param msg The UpdateMessage that was processed.
|
*/
|
void processUpdateDoneSynchronous(UpdateMsg msg)
|
{
|
/*
|
Warning: in synchronous mode, no way to tell the replay of an update went
|
wrong Just put null in processUpdateDone so that if assured replication
|
is used the ack is sent without error at replay flag.
|
*/
|
processUpdateDone(msg, null);
|
state.update(msg.getCSN());
|
}
|
|
/**
|
* Check if the domain is connected to a ReplicationServer.
|
*
|
* @return true if the server is connected, false if not.
|
*/
|
public boolean isConnected()
|
{
|
return broker != null && broker.isConnected();
|
}
|
|
/**
|
* Check if the domain has a connection error.
|
* A Connection error happens when the broker could not be created
|
* or when the broker could not find any ReplicationServer to connect to.
|
*
|
* @return true if the domain has a connection error.
|
*/
|
public boolean hasConnectionError()
|
{
|
return broker == null || broker.hasConnectionError();
|
}
|
|
/**
|
* Get the name of the replicationServer to which this domain is currently
|
* connected.
|
*
|
* @return the name of the replicationServer to which this domain
|
* is currently connected.
|
*/
|
public String getReplicationServer()
|
{
|
if (broker != null)
|
return broker.getReplicationServer();
|
return ReplicationBroker.NO_CONNECTED_SERVER;
|
}
|
|
/**
|
* Gets the number of updates sent in assured safe read mode.
|
* @return The number of updates sent in assured safe read mode.
|
*/
|
public int getAssuredSrSentUpdates()
|
{
|
return assuredSrSentUpdates.get();
|
}
|
|
/**
|
* Gets the number of updates sent in assured safe read mode that have been
|
* acknowledged without errors.
|
* @return The number of updates sent in assured safe read mode that have been
|
* acknowledged without errors.
|
*/
|
public int getAssuredSrAcknowledgedUpdates()
|
{
|
return assuredSrAcknowledgedUpdates.get();
|
}
|
|
/**
|
* Gets the number of updates sent in assured safe read mode that have not
|
* been acknowledged.
|
* @return The number of updates sent in assured safe read mode that have not
|
* been acknowledged.
|
*/
|
public int getAssuredSrNotAcknowledgedUpdates()
|
{
|
return assuredSrNotAcknowledgedUpdates.get();
|
}
|
|
/**
|
* Gets the number of updates sent in assured safe read mode that have not
|
* been acknowledged due to timeout error.
|
* @return The number of updates sent in assured safe read mode that have not
|
* been acknowledged due to timeout error.
|
*/
|
public int getAssuredSrTimeoutUpdates()
|
{
|
return assuredSrTimeoutUpdates.get();
|
}
|
|
/**
|
* Gets the number of updates sent in assured safe read mode that have not
|
* been acknowledged due to wrong status error.
|
* @return The number of updates sent in assured safe read mode that have not
|
* been acknowledged due to wrong status error.
|
*/
|
public int getAssuredSrWrongStatusUpdates()
|
{
|
return assuredSrWrongStatusUpdates.get();
|
}
|
|
/**
|
* Gets the number of updates sent in assured safe read mode that have not
|
* been acknowledged due to replay error.
|
* @return The number of updates sent in assured safe read mode that have not
|
* been acknowledged due to replay error.
|
*/
|
public int getAssuredSrReplayErrorUpdates()
|
{
|
return assuredSrReplayErrorUpdates.get();
|
}
|
|
/**
|
* Gets the number of updates sent in assured safe read mode that have not
|
* been acknowledged per server.
|
* @return A copy of the map that contains the number of updates sent in
|
* assured safe read mode that have not been acknowledged per server.
|
*/
|
public Map<Integer, Integer> getAssuredSrServerNotAcknowledgedUpdates()
|
{
|
synchronized(assuredSrServerNotAcknowledgedUpdates)
|
{
|
return new HashMap<Integer, Integer>(
|
assuredSrServerNotAcknowledgedUpdates);
|
}
|
}
|
|
/**
|
* Gets the number of updates received in assured safe read mode request.
|
* @return The number of updates received in assured safe read mode request.
|
*/
|
public int getAssuredSrReceivedUpdates()
|
{
|
return assuredSrReceivedUpdates.get();
|
}
|
|
/**
|
* Gets the number of updates received in assured safe read mode that we acked
|
* without error (no replay error).
|
* @return The number of updates received in assured safe read mode that we
|
* acked without error (no replay error).
|
*/
|
public int getAssuredSrReceivedUpdatesAcked()
|
{
|
return this.assuredSrReceivedUpdatesAcked.get();
|
}
|
|
/**
|
* Gets the number of updates received in assured safe read mode that we did
|
* not ack due to error (replay error).
|
* @return The number of updates received in assured safe read mode that we
|
* did not ack due to error (replay error).
|
*/
|
public int getAssuredSrReceivedUpdatesNotAcked()
|
{
|
return this.assuredSrReceivedUpdatesNotAcked.get();
|
}
|
|
/**
|
* Gets the number of updates sent in assured safe data mode.
|
* @return The number of updates sent in assured safe data mode.
|
*/
|
public int getAssuredSdSentUpdates()
|
{
|
return assuredSdSentUpdates.get();
|
}
|
|
/**
|
* Gets the number of updates sent in assured safe data mode that have been
|
* acknowledged without errors.
|
* @return The number of updates sent in assured safe data mode that have been
|
* acknowledged without errors.
|
*/
|
public int getAssuredSdAcknowledgedUpdates()
|
{
|
return assuredSdAcknowledgedUpdates.get();
|
}
|
|
/**
|
* Gets the number of updates sent in assured safe data mode that have not
|
* been acknowledged due to timeout error.
|
* @return The number of updates sent in assured safe data mode that have not
|
* been acknowledged due to timeout error.
|
*/
|
public int getAssuredSdTimeoutUpdates()
|
{
|
return assuredSdTimeoutUpdates.get();
|
}
|
|
/**
|
* Gets the number of updates sent in assured safe data mode that have not
|
* been acknowledged due to timeout error per server.
|
* @return A copy of the map that contains the number of updates sent in
|
* assured safe data mode that have not been acknowledged due to timeout
|
* error per server.
|
*/
|
public Map<Integer, Integer> getAssuredSdServerTimeoutUpdates()
|
{
|
synchronized(assuredSdServerTimeoutUpdates)
|
{
|
return new HashMap<Integer, Integer>(assuredSdServerTimeoutUpdates);
|
}
|
}
|
|
/**
|
* Gets the date of the last status change.
|
* @return The date of the last status change.
|
*/
|
public Date getLastStatusChangeDate()
|
{
|
return lastStatusChangeDate;
|
}
|
|
/**
|
* Resets the values of the monitoring counters.
|
*/
|
private void resetMonitoringCounters()
|
{
|
numProcessedUpdates = new AtomicInteger(0);
|
numRcvdUpdates = new AtomicInteger(0);
|
numSentUpdates = new AtomicInteger(0);
|
|
assuredSrSentUpdates = new AtomicInteger(0);
|
assuredSrAcknowledgedUpdates = new AtomicInteger(0);
|
assuredSrNotAcknowledgedUpdates = new AtomicInteger(0);
|
assuredSrTimeoutUpdates = new AtomicInteger(0);
|
assuredSrWrongStatusUpdates = new AtomicInteger(0);
|
assuredSrReplayErrorUpdates = new AtomicInteger(0);
|
synchronized (assuredSrServerNotAcknowledgedUpdates)
|
{
|
assuredSrServerNotAcknowledgedUpdates.clear();
|
}
|
assuredSrReceivedUpdates = new AtomicInteger(0);
|
assuredSrReceivedUpdatesAcked = new AtomicInteger(0);
|
assuredSrReceivedUpdatesNotAcked = new AtomicInteger(0);
|
assuredSdSentUpdates = new AtomicInteger(0);
|
assuredSdAcknowledgedUpdates = new AtomicInteger(0);
|
assuredSdTimeoutUpdates = new AtomicInteger(0);
|
synchronized (assuredSdServerTimeoutUpdates)
|
{
|
assuredSdServerTimeoutUpdates.clear();
|
}
|
}
|
|
/*
|
********** End of Monitoring Code **************
|
*/
|
|
/**
|
* Start the publish mechanism of the Replication Service.
|
* After this method has been called, the publish service can be used
|
* by calling the {@link #publish(UpdateMsg)} method.
|
*
|
* @param replicationServers The replication servers that should be used.
|
* @param window The window size of this replication domain.
|
* @param heartbeatInterval The heartbeatInterval that should be used
|
* to check the availability of the replication
|
* servers.
|
* @param changetimeHeartbeatInterval The interval used to send change
|
* time heartbeat to the replication server.
|
*
|
* @throws ConfigException If the DirectoryServer configuration was
|
* incorrect.
|
*/
|
public void startPublishService(Set<String> replicationServers, int window,
|
long heartbeatInterval, long changetimeHeartbeatInterval)
|
throws ConfigException
|
{
|
synchronized (sessionLock)
|
{
|
if (broker == null)
|
{
|
// create the broker object used to publish and receive changes
|
broker = new ReplicationBroker(
|
this, state, baseDN,
|
serverID, window,
|
getGenerationID(),
|
heartbeatInterval,
|
new ReplSessionSecurity(),
|
getGroupId(),
|
changetimeHeartbeatInterval);
|
|
broker.start(replicationServers);
|
}
|
}
|
}
|
|
/**
|
* Starts the receiver side of the Replication Service.
|
* <p>
|
* After this method has been called, the Replication Service will start
|
* calling the {@link #processUpdate(UpdateMsg, AtomicBoolean)}.
|
* <p>
|
* This method must be called once and must be called after the
|
* {@link #startPublishService(Collection, int, long, long)}.
|
*/
|
public void startListenService()
|
{
|
synchronized (sessionLock)
|
{
|
// Create the listener thread
|
listenerThread = new ListenerThread(this);
|
listenerThread.start();
|
}
|
}
|
|
/**
|
* Temporarily disable the Replication Service.
|
* The Replication Service can be enabled again using
|
* {@link #enableService()}.
|
* <p>
|
* It can be useful to disable the Replication Service when the
|
* repository where the replicated information is stored becomes
|
* temporarily unavailable and replicated updates can therefore not
|
* be replayed during a while. This method is not MT safe.
|
*/
|
public void disableService()
|
{
|
synchronized (sessionLock)
|
{
|
/*
|
Stop the broker first in order to prevent the listener from
|
reconnecting - see OPENDJ-457.
|
*/
|
if (broker != null)
|
{
|
broker.stop();
|
}
|
|
// Stop the listener thread
|
if (listenerThread != null)
|
{
|
listenerThread.shutdown();
|
listenerThread.waitForShutdown();
|
listenerThread = null;
|
}
|
}
|
}
|
|
/**
|
* Restart the Replication service after a {@link #disableService()}.
|
* <p>
|
* The Replication Service will restart from the point indicated by the
|
* {@link ServerState} that was given as a parameter to the
|
* {@link #startPublishService(Collection, int, long, long)}
|
* at startup time.
|
* If some data have changed in the repository during the period of time when
|
* the Replication Service was disabled, this {@link ServerState} should
|
* therefore be updated by the Replication Domain subclass before calling
|
* this method. This method is not MT safe.
|
*/
|
public void enableService()
|
{
|
synchronized (sessionLock)
|
{
|
broker.start();
|
|
// Create the listener thread
|
listenerThread = new ListenerThread(this);
|
listenerThread.start();
|
}
|
}
|
|
/**
|
* Definitively stops the Replication Service.
|
*/
|
public void stopDomain()
|
{
|
disableService();
|
domains.remove(baseDN);
|
}
|
|
/**
|
* Change some ReplicationDomain parameters.
|
*
|
* @param replicationServers The new set of Replication Servers that this
|
* domain should now use.
|
* @param windowSize The window size that this domain should use.
|
* @param heartbeatInterval The heartbeatInterval that this domain should
|
* use.
|
* @param groupId The new group id to use
|
*/
|
public void changeConfig(Set<String> replicationServers, int windowSize,
|
long heartbeatInterval, byte groupId)
|
{
|
this.groupId = groupId;
|
|
if (broker != null
|
&& broker.changeConfig(replicationServers, windowSize,
|
heartbeatInterval, groupId))
|
{
|
disableService();
|
enableService();
|
}
|
}
|
|
|
|
/**
|
* Applies a configuration change to the attributes which should be be
|
* included in the ECL.
|
*
|
* @param includeAttributes
|
* attributes to be included with all change records.
|
* @param includeAttributesForDeletes
|
* additional attributes to be included with delete change records.
|
*/
|
public void changeConfig(Set<String> includeAttributes,
|
Set<String> includeAttributesForDeletes)
|
{
|
if (setEclIncludes(serverID, includeAttributes, includeAttributesForDeletes)
|
&& broker != null)
|
{
|
disableService();
|
enableService();
|
}
|
}
|
|
|
|
/**
|
* This method should trigger an export of the replicated data.
|
* to the provided outputStream.
|
* When finished the outputStream should be flushed and closed.
|
*
|
* @param output The OutputStream where the export should
|
* be produced.
|
* @throws DirectoryException When needed.
|
*/
|
abstract protected void exportBackend(OutputStream output)
|
throws DirectoryException;
|
|
/**
|
* This method should trigger an import of the replicated data.
|
*
|
* @param input The InputStream from which
|
* the import should be reading entries.
|
*
|
* @throws DirectoryException When needed.
|
*/
|
abstract protected void importBackend(InputStream input)
|
throws DirectoryException;
|
|
/**
|
* This method should return the total number of objects in the
|
* replicated domain.
|
* This count will be used for reporting.
|
*
|
* @throws DirectoryException when needed.
|
*
|
* @return The number of objects in the replication domain.
|
*/
|
public abstract long countEntries() throws DirectoryException;
|
|
/**
|
* This method should handle the processing of {@link UpdateMsg} receive from
|
* remote replication entities.
|
* <p>
|
* This method will be called by a single thread and should therefore should
|
* not be blocking.
|
*
|
* @param updateMsg
|
* The {@link UpdateMsg} that was received.
|
* @param shutdown
|
* whether the server initiated shutdown
|
* @return A boolean indicating if the processing is completed at return time.
|
* If <code> true </code> is returned, no further processing is
|
* necessary. If <code> false </code> is returned, the subclass should
|
* call the method {@link #processUpdateDone(UpdateMsg, String)} and
|
* update the ServerState When this processing is complete.
|
*/
|
public abstract boolean processUpdate(UpdateMsg updateMsg,
|
AtomicBoolean shutdown);
|
|
/**
|
* This method must be called after each call to
|
* {@link #processUpdate(UpdateMsg, AtomicBoolean)} when the processing of the
|
* update is completed.
|
* <p>
|
* It is useful for implementation needing to process the update in an
|
* asynchronous way or using several threads, but must be called even by
|
* implementation doing it in a synchronous, single-threaded way.
|
*
|
* @param msg
|
* The UpdateMsg whose processing was completed.
|
* @param replayErrorMsg
|
* if not null, this means an error occurred during the replay of
|
* this update, and this is the matching human readable message
|
* describing the problem.
|
*/
|
public void processUpdateDone(UpdateMsg msg, String replayErrorMsg)
|
{
|
broker.updateWindowAfterReplay();
|
|
/*
|
Send an ack if it was requested and the group id is the same of the RS
|
one. Only Safe Read mode makes sense in DS for returning an ack.
|
*/
|
byte rsGroupId = broker.getRsGroupId();
|
if (msg.isAssured())
|
{
|
// Assured feature is supported starting from replication protocol V2
|
if (broker.getProtocolVersion() >=
|
ProtocolVersion.REPLICATION_PROTOCOL_V2)
|
{
|
AssuredMode msgAssuredMode = msg.getAssuredMode();
|
if (msgAssuredMode == AssuredMode.SAFE_READ_MODE)
|
{
|
if (rsGroupId == groupId)
|
{
|
// Send the ack
|
AckMsg ackMsg = new AckMsg(msg.getCSN());
|
if (replayErrorMsg != null)
|
{
|
// Mark the error in the ack
|
// -> replay error occurred
|
ackMsg.setHasReplayError(true);
|
// -> replay error occurred in our server
|
List<Integer> idList = new ArrayList<Integer>();
|
idList.add(serverID);
|
ackMsg.setFailedServers(idList);
|
}
|
broker.publish(ackMsg);
|
if (replayErrorMsg != null)
|
{
|
assuredSrReceivedUpdatesNotAcked.incrementAndGet();
|
} else
|
{
|
assuredSrReceivedUpdatesAcked.incrementAndGet();
|
}
|
}
|
} else if (assuredMode != AssuredMode.SAFE_DATA_MODE)
|
{
|
Message errorMsg = ERR_DS_UNKNOWN_ASSURED_MODE.get(
|
Integer.toString(serverID), msgAssuredMode.toString(),
|
getBaseDNString(), msg.toString());
|
logError(errorMsg);
|
}
|
// Nothing to do in Assured safe data mode, only RS ack updates.
|
}
|
}
|
|
incProcessedUpdates();
|
}
|
|
/**
|
* Prepare a message if it is to be sent in assured mode.
|
* If the assured mode is enabled, this method should be called before
|
* publish(UpdateMsg msg) method. This will configure the update accordingly
|
* before it is sent and will prepare the mechanism that will block until the
|
* matching ack is received. To wait for the ack after publish call, use
|
* the waitForAckIfAssuredEnabled() method.
|
* The expected typical usage in a service inheriting from this class is
|
* the following sequence:
|
* UpdateMsg msg = xxx;
|
* prepareWaitForAckIfAssuredEnabled(msg);
|
* publish(msg);
|
* waitForAckIfAssuredEnabled(msg);
|
*
|
* Note: prepareWaitForAckIfAssuredEnabled and waitForAckIfAssuredEnabled have
|
* no effect if assured replication is disabled.
|
* Note: this mechanism should not be used if using publish(byte[] msg)
|
* version as usage of these methods is already hidden inside.
|
*
|
* @param msg The update message to be sent soon.
|
*/
|
protected void prepareWaitForAckIfAssuredEnabled(UpdateMsg msg)
|
{
|
byte rsGroupId = broker.getRsGroupId();
|
/*
|
* If assured configured, set message accordingly to request an ack in the
|
* right assured mode.
|
* No ack requested for a RS with a different group id. Assured
|
* replication supported for the same locality, i.e: a topology working in
|
* the same
|
* geographical location). If we are connected to a RS which is not in our
|
* locality, no need to ask for an ack.
|
*/
|
if (assured && rsGroupId == groupId)
|
{
|
msg.setAssured(true);
|
msg.setAssuredMode(assuredMode);
|
if (assuredMode == AssuredMode.SAFE_DATA_MODE)
|
msg.setSafeDataLevel(assuredSdLevel);
|
|
// Add the assured message to the list of update that are waiting for acks
|
waitingAckMsgs.put(msg.getCSN(), msg);
|
}
|
}
|
|
/**
|
* Wait for the processing of an assured message after it has been sent, if
|
* assured replication is configured, otherwise, do nothing.
|
* The prepareWaitForAckIfAssuredEnabled method should have been called
|
* before, see its comment for the full picture.
|
*
|
* @param msg The UpdateMsg for which we are waiting for an ack.
|
* @throws TimeoutException When the configured timeout occurs waiting for the
|
* ack.
|
*/
|
protected void waitForAckIfAssuredEnabled(UpdateMsg msg)
|
throws TimeoutException
|
{
|
byte rsGroupId = broker.getRsGroupId();
|
|
// If assured mode configured, wait for acknowledgement for the just sent
|
// message
|
if (assured && rsGroupId == groupId)
|
{
|
// Increment assured replication monitoring counters
|
switch (assuredMode)
|
{
|
case SAFE_READ_MODE:
|
assuredSrSentUpdates.incrementAndGet();
|
break;
|
case SAFE_DATA_MODE:
|
assuredSdSentUpdates.incrementAndGet();
|
break;
|
default:
|
// Should not happen
|
}
|
} else
|
{
|
// Not assured or bad group id, return immediately
|
return;
|
}
|
|
// Wait for the ack to be received, timing out if necessary
|
long startTime = System.currentTimeMillis();
|
synchronized (msg)
|
{
|
CSN csn = msg.getCSN();
|
while (waitingAckMsgs.containsKey(csn))
|
{
|
try
|
{
|
/*
|
WARNING: this timeout may be difficult to optimize: too low, it
|
may use too much CPU, too high, it may penalize performance...
|
*/
|
msg.wait(10);
|
} catch (InterruptedException e)
|
{
|
if (debugEnabled())
|
{
|
TRACER.debugInfo("waitForAck method interrupted for replication " +
|
"baseDN: " + baseDN);
|
}
|
break;
|
}
|
// Timeout ?
|
if ( (System.currentTimeMillis() - startTime) >= assuredTimeout )
|
{
|
/*
|
Timeout occurred, be sure that ack is not being received and if so,
|
remove the update from the wait list, log the timeout error and
|
also update assured monitoring counters
|
*/
|
UpdateMsg update = waitingAckMsgs.remove(csn);
|
|
if (update != null)
|
{
|
// No luck, this is a real timeout
|
// Increment assured replication monitoring counters
|
switch (msg.getAssuredMode())
|
{
|
case SAFE_READ_MODE:
|
assuredSrNotAcknowledgedUpdates.incrementAndGet();
|
assuredSrTimeoutUpdates.incrementAndGet();
|
// Increment number of errors for our RS
|
updateAssuredErrorsByServer(
|
assuredSrServerNotAcknowledgedUpdates,
|
broker.getRsServerId());
|
break;
|
case SAFE_DATA_MODE:
|
assuredSdTimeoutUpdates.incrementAndGet();
|
// Increment number of errors for our RS
|
updateAssuredErrorsByServer(assuredSdServerTimeoutUpdates,
|
broker.getRsServerId());
|
break;
|
default:
|
// Should not happen
|
}
|
|
throw new TimeoutException("No ack received for message csn: "
|
+ csn + " and replication servceID: " + baseDN + " after "
|
+ assuredTimeout + " ms.");
|
} else
|
{
|
// Ack received just before timeout limit: we can exit
|
break;
|
}
|
}
|
}
|
}
|
}
|
|
/**
|
* Publish an {@link UpdateMsg} to the Replication Service.
|
* <p>
|
* The Replication Service will handle the delivery of this {@link UpdateMsg}
|
* to all the participants of this Replication Domain. These members will be
|
* receive this {@link UpdateMsg} through a call of the
|
* {@link #processUpdate(UpdateMsg, AtomicBoolean)} message.
|
*
|
* @param msg The UpdateMsg that should be pushed.
|
*/
|
public void publish(UpdateMsg msg)
|
{
|
// Publish the update
|
broker.publish(msg);
|
state.update(msg.getCSN());
|
numSentUpdates.incrementAndGet();
|
}
|
|
/**
|
* Publish information to the Replication Service (not assured mode).
|
*
|
* @param msg The byte array containing the information that should
|
* be sent to the remote entities.
|
*/
|
public void publish(byte[] msg)
|
{
|
UpdateMsg update;
|
synchronized (this)
|
{
|
update = new UpdateMsg(generator.newCSN(), msg);
|
/*
|
If assured replication is configured, this will prepare blocking
|
mechanism. If assured replication is disabled, this returns
|
immediately
|
*/
|
prepareWaitForAckIfAssuredEnabled(update);
|
|
publish(update);
|
}
|
|
try
|
{
|
/*
|
If assured replication is enabled, this will wait for the matching
|
ack or time out. If assured replication is disabled, this returns
|
immediately
|
*/
|
waitForAckIfAssuredEnabled(update);
|
} catch (TimeoutException ex)
|
{
|
// This exception may only be raised if assured replication is
|
// enabled
|
Message errorMsg = NOTE_DS_ACK_TIMEOUT.get(getBaseDNString(),
|
Long.toString(assuredTimeout), update.toString());
|
logError(errorMsg);
|
}
|
}
|
|
/**
|
* This method should return the generationID to use for this
|
* ReplicationDomain.
|
* This method can be called at any time after the ReplicationDomain
|
* has been started.
|
*
|
* @return The GenerationID.
|
*/
|
public abstract long getGenerationID();
|
|
/**
|
* Subclasses should use this method to add additional monitoring
|
* information in the ReplicationDomain.
|
*
|
* @return Additional monitoring attributes that will be added in the
|
* ReplicationDomain monitoring entry.
|
*/
|
public Collection<Attribute> getAdditionalMonitoring()
|
{
|
return new ArrayList<Attribute>();
|
}
|
|
/**
|
* Returns a boolean indicating if a total update import is currently
|
* in Progress.
|
*
|
* @return A boolean indicating if a total update import is currently
|
* in Progress.
|
*/
|
public boolean importInProgress()
|
{
|
return ieContext != null && ieContext.importInProgress;
|
}
|
|
/**
|
* Returns a boolean indicating if a total update export is currently
|
* in Progress.
|
*
|
* @return A boolean indicating if a total update export is currently
|
* in Progress.
|
*/
|
public boolean exportInProgress()
|
{
|
return ieContext != null && !ieContext.importInProgress;
|
}
|
|
/**
|
* Returns the number of entries still to be processed when a total update
|
* is in progress.
|
*
|
* @return The number of entries still to be processed when a total update
|
* is in progress.
|
*/
|
long getLeftEntryCount()
|
{
|
if (ieContext != null)
|
return ieContext.entryLeftCount;
|
return 0;
|
}
|
|
/**
|
* Returns the local address of this replication domain, or the empty string
|
* if it is not yet connected.
|
*
|
* @return The local address.
|
*/
|
String getLocalUrl()
|
{
|
final ReplicationBroker tmp = broker;
|
return tmp != null ? tmp.getLocalUrl() : "";
|
}
|
|
/**
|
* Returns the total number of entries to be processed when a total update
|
* is in progress.
|
*
|
* @return The total number of entries to be processed when a total update
|
* is in progress.
|
*/
|
long getTotalEntryCount()
|
{
|
if (ieContext != null)
|
return ieContext.entryCount;
|
return 0;
|
}
|
|
|
|
/**
|
* Set the attributes configured on a server to be included in the ECL.
|
*
|
* @param serverId
|
* Server where these attributes are configured.
|
* @param includeAttributes
|
* Attributes to be included with all change records, may include
|
* wild-cards.
|
* @param includeAttributesForDeletes
|
* Additional attributes to be included with delete change records,
|
* may include wild-cards.
|
* @return {@code true} if the set of attributes was modified.
|
*/
|
public boolean setEclIncludes(int serverId,
|
Set<String> includeAttributes,
|
Set<String> includeAttributesForDeletes)
|
{
|
boolean configurationChanged = false;
|
|
synchronized (eclIncludesLock)
|
{
|
Set<String> s1 = new HashSet<String>(includeAttributes);
|
|
// Combine all+delete attributes.
|
Set<String> s2 = new HashSet<String>(s1);
|
s2.addAll(includeAttributesForDeletes);
|
|
if (!s1.equals(eclIncludesByServer.get(serverId)))
|
{
|
configurationChanged = true;
|
eclIncludesByServer.put(serverId, Collections.unmodifiableSet(s1));
|
}
|
|
if (!s2.equals(eclIncludesForDeletesByServer.get(serverId)))
|
{
|
configurationChanged = true;
|
eclIncludesForDeletesByServer.put(serverId,
|
Collections.unmodifiableSet(s2));
|
}
|
|
// and rebuild the global list to be ready for usage
|
Set<String> s = new HashSet<String>();
|
for (Set<String> attributes : eclIncludesByServer.values())
|
{
|
s.addAll(attributes);
|
}
|
eclIncludesAllServers = Collections.unmodifiableSet(s);
|
|
s = new HashSet<String>();
|
for (Set<String> attributes : eclIncludesForDeletesByServer.values())
|
{
|
s.addAll(attributes);
|
}
|
eclIncludesForDeletesAllServers = Collections.unmodifiableSet(s);
|
}
|
|
return configurationChanged;
|
}
|
|
|
|
/**
|
* Get the attributes to include in each change for the ECL.
|
*
|
* @return The attributes to include in each change for the ECL.
|
*/
|
public Set<String> getEclIncludes()
|
{
|
synchronized (eclIncludesLock)
|
{
|
return eclIncludesAllServers;
|
}
|
}
|
|
|
|
/**
|
* Get the attributes to include in each delete change for the ECL.
|
*
|
* @return The attributes to include in each delete change for the ECL.
|
*/
|
public Set<String> getEclIncludesForDeletes()
|
{
|
synchronized (eclIncludesLock)
|
{
|
return eclIncludesForDeletesAllServers;
|
}
|
}
|
|
|
|
/**
|
* Get the attributes to include in each change for the ECL for a given
|
* serverId.
|
*
|
* @param serverId
|
* The serverId for which we want the include attributes.
|
* @return The attributes.
|
*/
|
public Set<String> getEclIncludes(int serverId)
|
{
|
synchronized (eclIncludesLock)
|
{
|
return eclIncludesByServer.get(serverId);
|
}
|
}
|
|
|
|
/**
|
* Get the attributes to include in each change for the ECL for a given
|
* serverId.
|
*
|
* @param serverId
|
* The serverId for which we want the include attributes.
|
* @return The attributes.
|
*/
|
public Set<String> getEclIncludesForDeletes(int serverId)
|
{
|
synchronized (eclIncludesLock)
|
{
|
return eclIncludesForDeletesByServer.get(serverId);
|
}
|
}
|
|
/**
|
* Returns the CSN of the last Change that was fully processed by this
|
* ReplicationDomain.
|
*
|
* @return The CSN of the last Change that was fully processed by this
|
* ReplicationDomain.
|
*/
|
public CSN getLastLocalChange()
|
{
|
return state.getCSN(serverID);
|
}
|
|
/** {@inheritDoc} */
|
@Override
|
public String toString()
|
{
|
return getClass().getSimpleName() + " " + this.baseDN + " " + this.serverID;
|
}
|
}
|