/*
|
* CDDL HEADER START
|
*
|
* The contents of this file are subject to the terms of the
|
* Common Development and Distribution License, Version 1.0 only
|
* (the "License"). You may not use this file except in compliance
|
* with the License.
|
*
|
* You can obtain a copy of the license at
|
* trunk/opends/resource/legal-notices/OpenDS.LICENSE
|
* or https://OpenDS.dev.java.net/OpenDS.LICENSE.
|
* See the License for the specific language governing permissions
|
* and limitations under the License.
|
*
|
* When distributing Covered Code, include this CDDL HEADER in each
|
* file and include the License file at
|
* trunk/opends/resource/legal-notices/OpenDS.LICENSE. If applicable,
|
* add the following below this CDDL HEADER, with the fields enclosed
|
* by brackets "[]" replaced with your own identifying information:
|
* Portions Copyright [yyyy] [name of copyright owner]
|
*
|
* CDDL HEADER END
|
*
|
*
|
* Copyright 2008-2010 Sun Microsystems, Inc.
|
* Portions Copyright 2011 ForgeRock AS
|
*/
|
package org.opends.server.replication.service;
|
|
import static org.opends.messages.ReplicationMessages.*;
|
import static org.opends.server.loggers.ErrorLogger.logError;
|
import static org.opends.server.loggers.debug.DebugLogger.debugEnabled;
|
import static org.opends.server.loggers.debug.DebugLogger.getTracer;
|
import static org.opends.server.replication.common.StatusMachine.*;
|
|
import java.io.BufferedOutputStream;
|
import java.io.IOException;
|
import java.io.InputStream;
|
import java.io.OutputStream;
|
import java.net.SocketTimeoutException;
|
import java.util.ArrayList;
|
import java.util.Collection;
|
import java.util.Date;
|
import java.util.HashMap;
|
import java.util.HashSet;
|
import java.util.List;
|
import java.util.Map;
|
import java.util.Set;
|
import java.util.SortedMap;
|
import java.util.TreeMap;
|
import java.util.concurrent.ConcurrentHashMap;
|
import java.util.concurrent.TimeoutException;
|
import java.util.concurrent.atomic.AtomicInteger;
|
|
import org.opends.messages.Category;
|
import org.opends.messages.Message;
|
import org.opends.messages.Severity;
|
import org.opends.server.api.DirectoryThread;
|
import org.opends.server.backends.task.Task;
|
import org.opends.server.config.ConfigException;
|
import org.opends.server.core.DirectoryServer;
|
import org.opends.server.loggers.debug.DebugTracer;
|
import org.opends.server.replication.common.AssuredMode;
|
import org.opends.server.replication.common.ChangeNumber;
|
import org.opends.server.replication.common.ChangeNumberGenerator;
|
import org.opends.server.replication.common.DSInfo;
|
import org.opends.server.replication.common.RSInfo;
|
import org.opends.server.replication.common.ServerState;
|
import org.opends.server.replication.common.ServerStatus;
|
import org.opends.server.replication.common.StatusMachine;
|
import org.opends.server.replication.common.StatusMachineEvent;
|
import org.opends.server.replication.protocol.AckMsg;
|
import org.opends.server.replication.protocol.ChangeStatusMsg;
|
import org.opends.server.replication.protocol.DoneMsg;
|
import org.opends.server.replication.protocol.EntryMsg;
|
import org.opends.server.replication.protocol.ErrorMsg;
|
import org.opends.server.replication.protocol.HeartbeatMsg;
|
import org.opends.server.replication.protocol.InitializeRcvAckMsg;
|
import org.opends.server.replication.protocol.InitializeRequestMsg;
|
import org.opends.server.replication.protocol.InitializeTargetMsg;
|
import org.opends.server.replication.protocol.ProtocolSession;
|
import org.opends.server.replication.protocol.ProtocolVersion;
|
import org.opends.server.replication.protocol.ReplSessionSecurity;
|
import org.opends.server.replication.protocol.ReplicationMsg;
|
import org.opends.server.replication.protocol.ResetGenerationIdMsg;
|
import org.opends.server.replication.protocol.RoutableMsg;
|
import org.opends.server.replication.protocol.TopologyMsg;
|
import org.opends.server.replication.protocol.UpdateMsg;
|
import org.opends.server.tasks.InitializeTargetTask;
|
import org.opends.server.tasks.InitializeTask;
|
import org.opends.server.types.Attribute;
|
import org.opends.server.types.DirectoryException;
|
import org.opends.server.types.ResultCode;
|
|
/**
|
* This class should be used as a base for Replication implementations.
|
* <p>
|
* It is intended that developer in need of a replication mechanism
|
* subclass this class with their own implementation.
|
* <p>
|
* The startup phase of the ReplicationDomain subclass,
|
* should read the list of replication servers from the configuration,
|
* instantiate a {@link ServerState} then start the publish service
|
* by calling
|
* {@link #startPublishService(Collection, int, long, long)}.
|
* At this point it can start calling the {@link #publish(UpdateMsg)}
|
* method if needed.
|
* <p>
|
* When the startup phase reach the point when the subclass is ready
|
* to handle updates the Replication Domain implementation should call the
|
* {@link #startListenService()} method.
|
* At this point a Listener thread is created on the Replication Service
|
* and which can start receiving updates.
|
* <p>
|
* When updates are received the Replication Service calls the
|
* {@link #processUpdate(UpdateMsg)} method.
|
* ReplicationDomain implementation should implement the appropriate code
|
* for replaying the update on the local repository.
|
* When fully done the subclass must call the
|
* {@link #processUpdateDone(UpdateMsg, String)} method.
|
* This allows to process the update asynchronously if necessary.
|
*
|
* <p>
|
* To propagate changes to other replica, a ReplicationDomain implementation
|
* must use the {@link #publish(UpdateMsg)} method.
|
* <p>
|
* If the Full Initialization process is needed then implementation
|
* for {@code importBackend(InputStream)} and
|
* {@code exportBackend(OutputStream)} must be
|
* provided.
|
* <p>
|
* Full Initialization of a replica can be triggered by LDAP clients
|
* by creating InitializeTasks or InitializeTargetTask.
|
* Full initialization can also by triggered from the ReplicationDomain
|
* implementation using methods {@link #initializeRemote(int)}
|
* or {@link #initializeFromRemote(int)}.
|
* <p>
|
* At shutdown time, the {@link #stopDomain()} method should be called to
|
* cleanly stop the replication service.
|
*/
|
public abstract class ReplicationDomain
|
{
|
/**
|
* Current status for this replicated domain.
|
*/
|
protected ServerStatus status = ServerStatus.NOT_CONNECTED_STATUS;
|
|
/**
|
* The tracer object for the debug logger.
|
*/
|
private static final DebugTracer TRACER = getTracer();
|
|
/**
|
* An identifier for the Replication Service.
|
* All Replication Domain using this identifier will be connected
|
* through the Replication Service.
|
*/
|
private final String serviceID;
|
|
/**
|
* The identifier of this Replication Domain inside the
|
* Replication Service.
|
* Each Domain must use a unique ServerID.
|
*/
|
private final int serverID;
|
|
/**
|
* The ReplicationBroker that is used by this ReplicationDomain to
|
* connect to the ReplicationService.
|
*/
|
protected ReplicationBroker broker = null;
|
|
/**
|
* This Map is used to store all outgoing assured messages in order
|
* to be able to correlate all the coming back acks to the original
|
* operation.
|
*/
|
private final SortedMap<ChangeNumber, UpdateMsg> waitingAckMsgs =
|
new TreeMap<ChangeNumber, UpdateMsg>();
|
|
|
/**
|
* The context related to an import or export being processed
|
* Null when none is being processed.
|
*/
|
protected IEContext ieContext = null;
|
|
/**
|
* The Thread waiting for incoming update messages for this domain and pushing
|
* them to the global incoming update message queue for later processing by
|
* replay threads.
|
*/
|
private ListenerThread listenerThread;
|
|
/**
|
* A Map used to store all the ReplicationDomains created on this server.
|
*/
|
private static Map<String, ReplicationDomain> domains =
|
new HashMap<String, ReplicationDomain>();
|
|
/**
|
* The Monitor in charge of replication monitoring.
|
*/
|
private ReplicationMonitor monitor;
|
|
/*
|
* Assured mode properties
|
*/
|
// Is assured mode enabled or not for this domain ?
|
private boolean assured = false;
|
// Assured sub mode (used when assured is true)
|
private AssuredMode assuredMode = AssuredMode.SAFE_DATA_MODE;
|
// Safe Data level (used when assuredMode is SAFE_DATA)
|
private byte assuredSdLevel = (byte)1;
|
// The timeout in ms that should be used, when waiting for assured acks
|
private long assuredTimeout = 2000;
|
|
// Group id
|
private byte groupId = (byte)1;
|
// Referrals urls to be published to other servers of the topology
|
// TODO: fill that with all currently opened urls if no urls configured
|
private final List<String> refUrls = new ArrayList<String>();
|
|
/**
|
* A set of counters used for Monitoring.
|
*/
|
private AtomicInteger numProcessedUpdates = new AtomicInteger(0);
|
private AtomicInteger numRcvdUpdates = new AtomicInteger(0);
|
private AtomicInteger numSentUpdates = new AtomicInteger(0);
|
|
/* Assured replication monitoring counters */
|
|
// Number of updates sent in Assured Mode, Safe Read
|
private AtomicInteger assuredSrSentUpdates = new AtomicInteger(0);
|
// Number of updates sent in Assured Mode, Safe Read, that have been
|
// successfully acknowledged
|
private AtomicInteger assuredSrAcknowledgedUpdates = new AtomicInteger(0);
|
// Number of updates sent in Assured Mode, Safe Read, that have not been
|
// successfully acknowledged (either because of timeout, wrong status or error
|
// at replay)
|
private AtomicInteger assuredSrNotAcknowledgedUpdates =
|
new AtomicInteger(0);
|
// Number of updates sent in Assured Mode, Safe Read, that have not been
|
// successfully acknowledged because of timeout
|
private AtomicInteger assuredSrTimeoutUpdates = new AtomicInteger(0);
|
// Number of updates sent in Assured Mode, Safe Read, that have not been
|
// successfully acknowledged because of wrong status
|
private AtomicInteger assuredSrWrongStatusUpdates = new AtomicInteger(0);
|
// Number of updates sent in Assured Mode, Safe Read, that have not been
|
// successfully acknowledged because of replay error
|
private AtomicInteger assuredSrReplayErrorUpdates = new AtomicInteger(0);
|
// Multiple values allowed: number of updates sent in Assured Mode, Safe Read,
|
// that have not been successfully acknowledged (either because of timeout,
|
// wrong status or error at replay) for a particular server (DS or RS). String
|
// format: <server id>:<number of failed updates>
|
private Map<Integer, Integer> assuredSrServerNotAcknowledgedUpdates =
|
new HashMap<Integer,Integer>();
|
// Number of updates received in Assured Mode, Safe Read request
|
private AtomicInteger assuredSrReceivedUpdates = new AtomicInteger(0);
|
// Number of updates received in Assured Mode, Safe Read request that we have
|
// acked without errors
|
private AtomicInteger assuredSrReceivedUpdatesAcked = new AtomicInteger(0);
|
// Number of updates received in Assured Mode, Safe Read request that we have
|
// acked with errors
|
private AtomicInteger assuredSrReceivedUpdatesNotAcked = new AtomicInteger(0);
|
// Number of updates sent in Assured Mode, Safe Data
|
private AtomicInteger assuredSdSentUpdates = new AtomicInteger(0);
|
// Number of updates sent in Assured Mode, Safe Data, that have been
|
// successfully acknowledged
|
private AtomicInteger assuredSdAcknowledgedUpdates = new AtomicInteger(0);
|
// Number of updates sent in Assured Mode, Safe Data, that have not been
|
// successfully acknowledged because of timeout
|
private AtomicInteger assuredSdTimeoutUpdates = new AtomicInteger(0);
|
// Multiple values allowed: number of updates sent in Assured Mode, Safe Data,
|
// that have not been successfully acknowledged because of timeout for a
|
// particular RS. String format: <server id>:<number of failed updates>
|
private Map<Integer, Integer> assuredSdServerTimeoutUpdates =
|
new HashMap<Integer,Integer>();
|
|
/**
|
* Window size used during initialization .. between
|
* - the initializer/exporter DS that listens/waits acknowledges and that
|
* slows down data msg publishing based on the slowest server
|
* - and each initialized/importer DS that publishes acknowledges each
|
* WINDOW/2 data msg received.
|
*/
|
protected int initWindow = 100;
|
|
/* Status related monitoring fields */
|
|
// Indicates the date when the status changed. This may be used to indicate
|
// the date the session with the current replication server started (when
|
// status is NORMAL for instance). All the above assured monitoring fields
|
// are also reset each time the status is changed
|
private Date lastStatusChangeDate = new Date();
|
|
/**
|
* The state maintained by the Concrete Class.
|
*/
|
private final ServerState state;
|
|
/**
|
* The generator that will be used to generate {@link ChangeNumber}
|
* for this domain.
|
*/
|
private final ChangeNumberGenerator generator;
|
|
private final Map<Integer, Set<String>> eclIncludeByServer =
|
new ConcurrentHashMap<Integer, Set<String>>();
|
Set<String> crossServersECLIncludes = new HashSet<String>();
|
|
/**
|
* An object used to protect the initialization of the underlying broker
|
* session of this ReplicationDomain.
|
*/
|
private final Object sessionLock = new Object();
|
|
/**
|
* Returns the {@link ChangeNumberGenerator} that will be used to
|
* generate {@link ChangeNumber} for this domain.
|
*
|
* @return The {@link ChangeNumberGenerator} that will be used to
|
* generate {@link ChangeNumber} for this domain.
|
*/
|
public ChangeNumberGenerator getGenerator()
|
{
|
return generator;
|
}
|
|
/**
|
* Creates a ReplicationDomain with the provided parameters.
|
*
|
* @param serviceID The identifier of the Replication Domain to which
|
* this object is participating.
|
* @param serverID The identifier of the server that is participating
|
* to the Replication Domain.
|
* This identifier should be different for each server that
|
* is participating to a given Replication Domain.
|
* @param initWindow Window used during initialization.
|
*/
|
public ReplicationDomain(String serviceID, int serverID,int initWindow)
|
{
|
this.serviceID = serviceID;
|
this.serverID = serverID;
|
this.initWindow = initWindow;
|
this.state = new ServerState();
|
this.generator = new ChangeNumberGenerator(serverID, state);
|
|
domains.put(serviceID, this);
|
}
|
|
/**
|
* Creates a ReplicationDomain with the provided parameters.
|
*
|
* @param serviceID The identifier of the Replication Domain to which
|
* this object is participating.
|
* @param serverID The identifier of the server that is participating
|
* to the Replication Domain.
|
* This identifier should be different for each server that
|
* is participating to a given Replication Domain.
|
*/
|
public ReplicationDomain(String serviceID, int serverID)
|
{
|
this.serviceID = serviceID;
|
this.serverID = serverID;
|
this.state = new ServerState();
|
this.generator = new ChangeNumberGenerator(serverID, state);
|
|
domains.put(serviceID, this);
|
}
|
|
/**
|
* Creates a ReplicationDomain with the provided parameters.
|
* (for unit test purpose only)
|
*
|
* @param serviceID The identifier of the Replication Domain to which
|
* this object is participating.
|
* @param serverID The identifier of the server that is participating
|
* to the Replication Domain.
|
* This identifier should be different for each server that
|
* is participating to a given Replication Domain.
|
* @param serverState The serverState to use
|
*/
|
public ReplicationDomain(String serviceID, int serverID,
|
ServerState serverState)
|
{
|
this.serviceID = serviceID;
|
this.serverID = serverID;
|
this.state = serverState;
|
this.generator = new ChangeNumberGenerator(serverID, state);
|
|
domains.put(serviceID, this);
|
}
|
|
/**
|
* Set the initial status of the domain and perform necessary initializations.
|
* This method will be called by the Broker each time the ReplicationBroker
|
* establish a new session to a Replication Server.
|
*
|
* Implementations may override this method when they need to perform
|
* additional computing after session establishment.
|
* The default implementation should be sufficient for ReplicationDomains
|
* that don't need to perform additional computing.
|
*
|
* @param initStatus The status to enter the state machine with.
|
* @param replicationServerState The ServerState of the ReplicationServer
|
* with which the session was established.
|
* @param generationID The current generationID of the
|
* ReplicationServer with which the session
|
* was established.
|
* @param session The ProtocolSession that is currently used.
|
*/
|
public void sessionInitiated(
|
ServerStatus initStatus,
|
ServerState replicationServerState,
|
long generationID,
|
ProtocolSession session)
|
{
|
// Sanity check: is it a valid initial status?
|
if (!isValidInitialStatus(initStatus))
|
{
|
Message msg = ERR_DS_INVALID_INIT_STATUS.get(initStatus.toString(),
|
serviceID, Integer.toString(serverID));
|
logError(msg);
|
} else
|
{
|
status = initStatus;
|
}
|
generator.adjust(state);
|
generator.adjust(replicationServerState);
|
}
|
|
/**
|
* Processes an incoming ChangeStatusMsg. Compute new status according to
|
* given order. Then update domain for being compliant with new status
|
* definition.
|
* @param csMsg The received status message
|
*/
|
private void receiveChangeStatus(ChangeStatusMsg csMsg)
|
{
|
if (debugEnabled())
|
TRACER.debugInfo("Replication domain " + serviceID +
|
" received change status message:\n" + csMsg);
|
|
ServerStatus reqStatus = csMsg.getRequestedStatus();
|
|
// Translate requested status to a state machine event
|
StatusMachineEvent event = StatusMachineEvent.statusToEvent(reqStatus);
|
if (event == StatusMachineEvent.INVALID_EVENT)
|
{
|
Message msg = ERR_DS_INVALID_REQUESTED_STATUS.get(reqStatus.toString(),
|
serviceID, Integer.toString(serverID));
|
logError(msg);
|
return;
|
}
|
|
// Set the new status to the requested one
|
setNewStatus(event);
|
}
|
|
/**
|
* Called when first connection or disconnection detected.
|
*/
|
void toNotConnectedStatus()
|
{
|
// Go into not connected status
|
setNewStatus(StatusMachineEvent.TO_NOT_CONNECTED_STATUS_EVENT);
|
}
|
|
/**
|
* Perform whatever actions are needed to apply properties for being
|
* compliant with new status. Must be called in synchronized section for
|
* status. The new status is already set in status variable.
|
*/
|
private void updateDomainForNewStatus()
|
{
|
switch (status)
|
{
|
case NOT_CONNECTED_STATUS:
|
break;
|
case NORMAL_STATUS:
|
break;
|
case DEGRADED_STATUS:
|
break;
|
case FULL_UPDATE_STATUS:
|
// Signal RS we just entered the full update status
|
broker.signalStatusChange(status);
|
break;
|
case BAD_GEN_ID_STATUS:
|
break;
|
default:
|
if (debugEnabled())
|
TRACER.debugInfo("updateDomainForNewStatus: unexpected status: " +
|
status);
|
}
|
}
|
|
/**
|
* Gets the status for this domain.
|
* @return The status for this domain.
|
*/
|
public ServerStatus getStatus()
|
{
|
return status;
|
}
|
|
/**
|
* Gets the identifier of this domain.
|
*
|
* @return The identifier for this domain.
|
*/
|
public String getServiceID()
|
{
|
return serviceID;
|
}
|
|
/**
|
* Get the server ID.
|
* @return The server ID.
|
*/
|
public int getServerId()
|
{
|
return serverID;
|
}
|
|
/**
|
* Tells if assured replication is enabled for this domain.
|
* @return True if assured replication is enabled for this domain.
|
*/
|
public boolean isAssured()
|
{
|
return assured;
|
}
|
|
/**
|
* Gives the mode for the assured replication of the domain.
|
* @return The mode for the assured replication of the domain.
|
*/
|
public AssuredMode getAssuredMode()
|
{
|
return assuredMode;
|
}
|
|
/**
|
* Gives the assured level of the replication of the domain.
|
* @return The assured level of the replication of the domain.
|
*/
|
public byte getAssuredSdLevel()
|
{
|
return assuredSdLevel;
|
}
|
|
/**
|
* Gives the assured timeout of the replication of the domain (in ms).
|
* @return The assured timeout of the replication of the domain.
|
*/
|
public long getAssuredTimeout()
|
{
|
return assuredTimeout;
|
}
|
|
/**
|
* Gets the group id for this domain.
|
* @return The group id for this domain.
|
*/
|
public byte getGroupId()
|
{
|
return groupId;
|
}
|
|
/**
|
* Gets the referrals URLs this domain publishes.
|
* @return The referrals URLs this domain publishes.
|
*/
|
public List<String> getRefUrls()
|
{
|
return refUrls;
|
}
|
|
/**
|
* Gets the info for Replicas in the topology (except us).
|
* @return The info for Replicas in the topology (except us)
|
*/
|
public List<DSInfo> getReplicasList()
|
{
|
return broker.getDsList();
|
}
|
|
/**
|
* Returns informations about the DS server related to the provided serverId.
|
* based on the TopologyMsg we received when the remote replica connected or
|
* disconnected. Return null when no server with the provided serverId is
|
* connected.
|
*
|
* @param serverId The provided serverId of the remote replica
|
* @return the info related to this remote server if it is connected,
|
* null is the server is NOT connected.
|
*/
|
public DSInfo isRemoteDSConnected(int serverId)
|
{
|
for (DSInfo remoteDS : getReplicasList())
|
if (remoteDS.getDsId() == serverId)
|
return remoteDS;
|
return null;
|
}
|
|
/**
|
* Gets the States of all the Replicas currently in the
|
* Topology.
|
* When this method is called, a Monitoring message will be sent
|
* to the Replication Server to which this domain is currently connected
|
* so that it computes a table containing information about
|
* all Directory Servers in the topology.
|
* This Computation involves communications will all the servers
|
* currently connected and
|
*
|
* @return The States of all Replicas in the topology (except us)
|
*/
|
public Map<Integer, ServerState> getReplicaStates()
|
{
|
return broker.getReplicaStates();
|
}
|
|
/**
|
* Gets the info for RSs in the topology (except the one we are connected
|
* to).
|
* @return The info for RSs in the topology (except the one we are connected
|
* to)
|
*/
|
public List<RSInfo> getRsList()
|
{
|
return broker.getRsList();
|
}
|
|
|
/**
|
* Gets the server ID of the Replication Server to which the domain
|
* is currently connected.
|
*
|
* @return The server ID of the Replication Server to which the domain
|
* is currently connected.
|
*/
|
public int getRsServerId()
|
{
|
return broker.getRsServerId();
|
}
|
|
/**
|
* Increment the number of processed updates.
|
*/
|
private void incProcessedUpdates()
|
{
|
numProcessedUpdates.incrementAndGet();
|
}
|
|
/**
|
* get the number of updates replayed by the replication.
|
*
|
* @return The number of updates replayed by the replication
|
*/
|
int getNumProcessedUpdates()
|
{
|
if (numProcessedUpdates != null)
|
return numProcessedUpdates.get();
|
else
|
return 0;
|
}
|
|
/**
|
* get the number of updates received by the replication plugin.
|
*
|
* @return the number of updates received
|
*/
|
int getNumRcvdUpdates()
|
{
|
if (numRcvdUpdates != null)
|
return numRcvdUpdates.get();
|
else
|
return 0;
|
}
|
|
/**
|
* Get the number of updates sent by the replication plugin.
|
*
|
* @return the number of updates sent
|
*/
|
int getNumSentUpdates()
|
{
|
if (numSentUpdates != null)
|
return numSentUpdates.get();
|
else
|
return 0;
|
}
|
|
/**
|
* Set the list of Referrals that should be returned when an
|
* operation needs to be redirected to this server.
|
*
|
* @param referralsUrl The list of referrals.
|
*/
|
public void setURLs(Set<String> referralsUrl)
|
{
|
for (String url : referralsUrl)
|
this.refUrls.add(url);
|
}
|
|
/**
|
* Set the timeout of the assured replication.
|
*
|
* @param assuredTimeout the timeout of the assured replication.
|
*/
|
public void setAssuredTimeout(long assuredTimeout)
|
{
|
this.assuredTimeout = assuredTimeout;
|
}
|
|
/**
|
* Sets the groupID.
|
*
|
* @param groupId The groupID.
|
*/
|
public void setGroupId(byte groupId)
|
{
|
this.groupId = groupId;
|
}
|
|
/**
|
* Sets the level of assured replication.
|
*
|
* @param assuredSdLevel The level of assured replication.
|
*/
|
public void setAssuredSdLevel(byte assuredSdLevel)
|
{
|
this.assuredSdLevel = assuredSdLevel;
|
}
|
|
/**
|
* Sets the assured replication mode.
|
*
|
* @param dataMode The assured replication mode.
|
*/
|
public void setAssuredMode(AssuredMode dataMode)
|
{
|
this.assuredMode = dataMode;
|
}
|
|
/**
|
* Sets assured replication.
|
*
|
* @param assured A boolean indicating if assured replication should be used.
|
*/
|
public void setAssured(boolean assured)
|
{
|
this.assured = assured;
|
}
|
|
/**
|
* Receives an update message from the replicationServer.
|
* The other types of messages are processed in an opaque way for the caller.
|
* Also responsible for updating the list of pending changes
|
* @return the received message - null if none
|
*/
|
UpdateMsg receive()
|
{
|
UpdateMsg update = null;
|
|
while (update == null)
|
{
|
InitializeRequestMsg initReqMsg = null;
|
ReplicationMsg msg;
|
try
|
{
|
msg = broker.receive(true, true, false);
|
if (msg == null)
|
{
|
// The server is in the shutdown process
|
return null;
|
}
|
|
if (debugEnabled())
|
if (!(msg instanceof HeartbeatMsg))
|
TRACER.debugVerbose("Message received <" + msg + ">");
|
|
if (msg instanceof AckMsg)
|
{
|
AckMsg ack = (AckMsg) msg;
|
receiveAck(ack);
|
}
|
else if (msg instanceof InitializeRequestMsg)
|
{
|
// Another server requests us to provide entries
|
// for a total update
|
initReqMsg = (InitializeRequestMsg)msg;
|
}
|
else if (msg instanceof InitializeTargetMsg)
|
{
|
// Another server is exporting its entries to us
|
InitializeTargetMsg initTargetMsg = (InitializeTargetMsg) msg;
|
|
// This must be done while we are still holding the
|
// broker lock because we are now going to receive a
|
// bunch of entries from the remote server and we
|
// want the import thread to catch them and
|
// not the ListenerThread.
|
initialize(initTargetMsg, initTargetMsg.getSenderID());
|
}
|
else if (msg instanceof ErrorMsg)
|
{
|
ErrorMsg errorMsg = (ErrorMsg)msg;
|
if (ieContext != null)
|
{
|
// This is an error termination for the 2 following cases :
|
// - either during an export
|
// - or before an import really started
|
// For example, when we publish a request and the
|
// replicationServer did not find the import source.
|
//
|
// A remote error during the import will be received in the
|
// receiveEntryBytes() method.
|
//
|
if (debugEnabled())
|
TRACER.debugInfo(
|
"[IE] processErrorMsg:" + this.serverID +
|
" serviceID: " + this.serviceID +
|
" Error Msg received: " + errorMsg);
|
|
if (errorMsg.getCreationTime() > ieContext.startTime)
|
{
|
// consider only ErrorMsg that relate to the current import/export
|
processErrorMsg(errorMsg);
|
}
|
else
|
{
|
// Simply log - happen when the ErrorMsg relates to a previous
|
// attempt of initialization while we have started a new one
|
// on this side.
|
logError(ERR_ERROR_MSG_RECEIVED.get(errorMsg.getDetails()));
|
}
|
}
|
else
|
{
|
// Simply log - happen if import/export has been terminated
|
// on our side before receiving this ErrorMsg.
|
logError(ERR_ERROR_MSG_RECEIVED.get(errorMsg.getDetails()));
|
}
|
}
|
else if (msg instanceof ChangeStatusMsg)
|
{
|
ChangeStatusMsg csMsg = (ChangeStatusMsg)msg;
|
receiveChangeStatus(csMsg);
|
}
|
else if (msg instanceof UpdateMsg)
|
{
|
update = (UpdateMsg) msg;
|
generator.adjust(update.getChangeNumber());
|
}
|
else if (msg instanceof InitializeRcvAckMsg)
|
{
|
if (ieContext != null)
|
{
|
InitializeRcvAckMsg ackMsg = (InitializeRcvAckMsg) msg;
|
ieContext.setAckVal(ackMsg.getSenderID(), ackMsg.getNumAck());
|
}
|
// Trash this msg When no input/export is running/should never happen
|
}
|
}
|
catch (SocketTimeoutException e)
|
{
|
// just retry
|
}
|
// Test if we have received and export request message and
|
// if that's the case handle it now.
|
// This must be done outside of the portion of code protected
|
// by the broker lock so that we keep receiveing update
|
// when we are doing and export and so that a possible
|
// closure of the socket happening when we are publishing the
|
// entries to the remote can be handled by the other
|
// replay thread when they call this method and therefore the
|
// broker.receive() method.
|
if (initReqMsg != null)
|
{
|
// Do this work in a thread to allow replay thread continue working
|
ExportThread exportThread = new ExportThread(
|
initReqMsg.getSenderID(), initReqMsg.getInitWindow());
|
exportThread.start();
|
}
|
}
|
|
numRcvdUpdates.incrementAndGet();
|
byte rsGroupId = broker.getRsGroupId();
|
if ( update.isAssured() && (update.getAssuredMode() ==
|
AssuredMode.SAFE_READ_MODE) && (rsGroupId == groupId) )
|
{
|
assuredSrReceivedUpdates.incrementAndGet();
|
}
|
return update;
|
}
|
|
/**
|
* Updates the passed monitoring list of errors received for assured messages
|
* (safe data or safe read, depending of the passed list to update) for a
|
* particular server in the list. This increments the counter of error for the
|
* passed server, or creates an initial value of 1 error for it if the server
|
* is not yet present in the map.
|
* @param errorList
|
* @param sid
|
*/
|
private void updateAssuredErrorsByServer(Map<Integer,Integer> errorsByServer,
|
Integer sid)
|
{
|
synchronized (errorsByServer)
|
{
|
Integer serverErrCount = errorsByServer.get(sid);
|
if (serverErrCount == null)
|
{
|
// Server not present in list, create an entry with an
|
// initial number of errors set to 1
|
errorsByServer.put(sid, 1);
|
} else
|
{
|
// Server already present in list, just increment number of
|
// errors for the server
|
int val = serverErrCount.intValue();
|
val++;
|
errorsByServer.put(sid, val);
|
}
|
}
|
}
|
|
/**
|
* Do the necessary processing when an AckMsg is received.
|
*
|
* @param ack The AckMsg that was received.
|
*/
|
private void receiveAck(AckMsg ack)
|
{
|
UpdateMsg update;
|
ChangeNumber changeNumber = ack.getChangeNumber();
|
|
// Remove the message for pending ack list (this may already make the thread
|
// that is waiting for the ack be aware of its reception)
|
synchronized (waitingAckMsgs)
|
{
|
update = waitingAckMsgs.remove(changeNumber);
|
}
|
|
// Signal waiting thread ack has been received
|
if (update != null)
|
{
|
synchronized (update)
|
{
|
update.notify();
|
}
|
|
// Analyze status of embedded in the ack to see if everything went well
|
boolean hasTimeout = ack.hasTimeout();
|
boolean hasReplayErrors = ack.hasReplayError();
|
boolean hasWrongStatus = ack.hasWrongStatus();
|
|
AssuredMode updateAssuredMode = update.getAssuredMode();
|
|
if ( hasTimeout || hasReplayErrors || hasWrongStatus)
|
{
|
// Some problems detected: message not correclty reached every requested
|
// servers. Log problem
|
Message errorMsg = NOTE_DS_RECEIVED_ACK_ERROR.get(
|
serviceID, Integer.toString(serverID),
|
update.toString(), ack.errorsToString());
|
logError(errorMsg);
|
|
List<Integer> failedServers = ack.getFailedServers();
|
|
// Increment assured replication monitoring counters
|
switch (updateAssuredMode)
|
{
|
case SAFE_READ_MODE:
|
assuredSrNotAcknowledgedUpdates.incrementAndGet();
|
if (hasTimeout)
|
assuredSrTimeoutUpdates.incrementAndGet();
|
if (hasReplayErrors)
|
assuredSrReplayErrorUpdates.incrementAndGet();
|
if (hasWrongStatus)
|
assuredSrWrongStatusUpdates.incrementAndGet();
|
if (failedServers != null) // This should always be the case !
|
{
|
for(Integer sid : failedServers)
|
{
|
updateAssuredErrorsByServer(
|
assuredSrServerNotAcknowledgedUpdates, sid);
|
}
|
}
|
break;
|
case SAFE_DATA_MODE:
|
// The only possible cause of ack error in safe data mode is timeout
|
if (hasTimeout) // So should always be the case
|
assuredSdTimeoutUpdates.incrementAndGet();
|
if (failedServers != null) // This should always be the case !
|
{
|
for(Integer sid : failedServers)
|
{
|
updateAssuredErrorsByServer(
|
assuredSdServerTimeoutUpdates, sid);
|
}
|
}
|
break;
|
default:
|
// Should not happen
|
}
|
} else
|
{
|
// Update has been acknowledged without errors
|
// Increment assured replication monitoring counters
|
switch (updateAssuredMode)
|
{
|
case SAFE_READ_MODE:
|
assuredSrAcknowledgedUpdates.incrementAndGet();
|
break;
|
case SAFE_DATA_MODE:
|
assuredSdAcknowledgedUpdates.incrementAndGet();
|
break;
|
default:
|
// Should not happen
|
}
|
}
|
}
|
}
|
|
/**
|
* Retrieves a replication domain based on the baseDn.
|
*
|
* @param serviceID The identifier of the domain to retrieve.
|
*
|
* @return The domain retrieved.
|
*
|
* @throws DirectoryException When an error occurred or no domain
|
* match the provided baseDn.
|
*/
|
static ReplicationDomain retrievesReplicationDomain(String serviceID)
|
throws DirectoryException
|
{
|
ReplicationDomain replicationDomain = domains.get(serviceID);
|
if (replicationDomain == null)
|
{
|
throw new DirectoryException(ResultCode.OTHER,
|
ERR_NO_MATCHING_DOMAIN.get(serviceID));
|
}
|
return replicationDomain;
|
}
|
|
/*
|
* After this point the code is related to the Total Update.
|
*/
|
|
/**
|
* This thread is launched when we want to export data to another server.
|
*
|
* When a task is created locally (so this local server is the initiator)
|
* of the export (Exemple: dsreplication initialize-all),
|
* this thread is NOT used but the task thread is running the export instead).
|
*/
|
private class ExportThread extends DirectoryThread
|
{
|
// Id of server that will be initialized
|
private final int serverToInitialize;
|
private final int initWindow;
|
|
|
|
/**
|
* Constructor for the ExportThread.
|
*
|
* @param serverToInitialize
|
* serverId of server that will receive entries
|
* @param initWindow
|
* The value of the initialization window for flow control between
|
* the importer and the exporter.
|
*/
|
public ExportThread(int serverToInitialize, int initWindow)
|
{
|
super("Export thread from serverId=" + serverID + " to serverId="
|
+ serverToInitialize);
|
this.serverToInitialize = serverToInitialize;
|
this.initWindow = initWindow;
|
}
|
|
|
|
/**
|
* Run method for this class.
|
*/
|
@Override
|
public void run()
|
{
|
if (debugEnabled())
|
TRACER.debugInfo("[IE] starting " + this.getName());
|
try
|
{
|
initializeRemote(serverToInitialize, serverToInitialize, null,
|
initWindow);
|
} catch (DirectoryException de)
|
{
|
// An error message has been sent to the peer
|
// This server is not the initiator of the export so there is
|
// nothing more to do locally.
|
}
|
|
if (debugEnabled())
|
TRACER.debugInfo("[IE] ending " + this.getName());
|
}
|
}
|
|
/**
|
* This class contain the context related to an import or export
|
* launched on the domain.
|
*/
|
protected class IEContext
|
{
|
// The private task that initiated the operation.
|
Task initializeTask;
|
// The destination in the case of an export
|
int exportTarget = RoutableMsg.UNKNOWN_SERVER;
|
// The source in the case of an import
|
int importSource = RoutableMsg.UNKNOWN_SERVER;
|
|
// The total entry count expected to be processed
|
long entryCount = 0;
|
// The count for the entry not yet processed
|
long entryLeftCount = 0;
|
|
// Exception raised during the initialization.
|
DirectoryException exception = null;
|
|
// Whether the context is related to an import or an export.
|
boolean importInProgress;
|
|
// Current counter of messages exchanged during the initialization
|
int msgCnt = 0;
|
|
// Number of connections lost when we start the initialization.
|
// Will help counting connections lost during initialization,
|
int initNumLostConnections = 0;
|
|
// Request message sent when this server has the initializeFromRemote task.
|
InitializeRequestMsg initReqMsgSent = null;
|
|
// Start time of the initialization process. ErrorMsg timestamped
|
// before thi startTime will be ignored.
|
long startTime;
|
|
// List fo replicas (DS) connected to the topology when
|
// initialization started.
|
Set<Integer> startList = new HashSet<Integer>(0);
|
|
// List fo replicas (DS) with a failure (disconnected from the topology)
|
// since the initialization started.
|
Set<Integer> failureList = new HashSet<Integer>(0);
|
|
// Flow control during initialization
|
// - for each remote server, counter of messages received
|
private final HashMap<Integer, Integer> ackVals =
|
new HashMap<Integer, Integer>();
|
// - serverId of the slowest server (the one with the smallest non null
|
// counter)
|
private int slowestServerId = -1;
|
|
short exporterProtocolVersion = -1;
|
|
// Window used during this initialization
|
int initWindow;
|
|
// Number of attempt already done for this initialization
|
short attemptCnt;
|
|
/**
|
* Creates a new IEContext.
|
*
|
* @param importInProgress true if the IEContext will be used
|
* for and import, false if the IEContext
|
* will be used for and export.
|
*/
|
public IEContext(boolean importInProgress)
|
{
|
this.importInProgress = importInProgress;
|
this.startTime = System.currentTimeMillis();
|
this.attemptCnt = 0;
|
|
}
|
|
/**
|
* Initializes the import/export counters with the provider value.
|
* @param total Total number of entries to be processed.
|
* @throws DirectoryException if an error occurred.
|
*/
|
private void initializeCounters(long total)
|
throws DirectoryException
|
{
|
entryCount = total;
|
entryLeftCount = total;
|
|
if (initializeTask != null)
|
{
|
if (initializeTask instanceof InitializeTask)
|
{
|
((InitializeTask)initializeTask).setTotal(entryCount);
|
((InitializeTask)initializeTask).setLeft(entryCount);
|
}
|
else if (initializeTask instanceof InitializeTargetTask)
|
{
|
((InitializeTargetTask)initializeTask).setTotal(entryCount);
|
((InitializeTargetTask)initializeTask).setLeft(entryCount);
|
}
|
}
|
}
|
|
/**
|
* Update the counters of the task for each entry processed during
|
* an import or export.
|
* @throws DirectoryException if an error occurred.
|
*/
|
public void updateCounters()
|
throws DirectoryException
|
{
|
entryLeftCount--;
|
|
if (initializeTask != null)
|
{
|
if (initializeTask instanceof InitializeTask)
|
{
|
((InitializeTask)initializeTask).setLeft(entryLeftCount);
|
}
|
else if (initializeTask instanceof InitializeTargetTask)
|
{
|
((InitializeTargetTask)initializeTask).setLeft(entryLeftCount);
|
}
|
}
|
}
|
|
/**
|
* Update the counters of the task for each entry processed during
|
* an import or export.
|
*
|
* @param entriesDone The number of entries that were processed
|
* since the last time this method was called.
|
*
|
* @throws DirectoryException if an error occurred.
|
*/
|
public void updateCounters(int entriesDone)
|
throws DirectoryException
|
{
|
entryLeftCount -= entriesDone;
|
|
if (initializeTask != null)
|
{
|
if (initializeTask instanceof InitializeTask)
|
{
|
((InitializeTask)initializeTask).setLeft(entryLeftCount);
|
}
|
else if (initializeTask instanceof InitializeTargetTask)
|
{
|
((InitializeTargetTask)initializeTask).setLeft(entryLeftCount);
|
}
|
}
|
}
|
|
/**
|
* {@inheritDoc}
|
*/
|
@Override
|
public String toString()
|
{
|
return new String("[ Entry count=" + this.entryCount +
|
", Entry left count=" + this.entryLeftCount + "]");
|
}
|
|
/**
|
* Gets the server id of the exporting server.
|
* @return the server id of the exporting server.
|
*/
|
public int getExportTarget()
|
{
|
return exportTarget;
|
}
|
|
/**
|
* Gets the server id of the importing server.
|
* @return the server id of the importing server.
|
*/
|
public int getImportSource()
|
{
|
return importSource;
|
}
|
|
/**
|
* Get the exception that occurred during the import/export.
|
* @return the exception that occurred during the import/export.
|
*/
|
public DirectoryException getException()
|
{
|
return exception;
|
}
|
|
/**
|
* Set the exception that occurred during the import/export.
|
* @param exception the exception that occurred during the import/export.
|
*/
|
public void setException(DirectoryException exception)
|
{
|
this.exception = exception;
|
}
|
|
/**
|
* Set the id of the EntryMsg acknowledged from a receiver (importer)server.
|
* (updated via the listener thread)
|
* @param serverId serverId of the acknowledger/receiver/importer server.
|
* @param numAck id of the message received.
|
*/
|
public void setAckVal(int serverId, int numAck)
|
{
|
if (debugEnabled())
|
TRACER.debugInfo("[IE] setAckVal[" + serverId + "]=" + numAck);
|
|
this.ackVals.put(serverId, numAck);
|
|
// Recompute the server with the minAck returned,means the slowest server.
|
slowestServerId = serverId;
|
for (Integer sid : ieContext.ackVals.keySet())
|
if (this.ackVals.get(sid) < this.ackVals.get(slowestServerId))
|
slowestServerId = sid;
|
}
|
|
/**
|
* Returns the serverId of the server that acknowledged the smallest
|
* EntryMsg id.
|
* @return serverId of the server with latest acknowledge.
|
* 0 when no ack has been received yet.
|
*/
|
public int getSlowestServer()
|
{
|
if (debugEnabled())
|
TRACER.debugInfo("[IE] getSlowestServer" + slowestServerId
|
+ " " + this.ackVals.get(slowestServerId));
|
|
return this.slowestServerId;
|
}
|
}
|
/**
|
* Verifies that the given string represents a valid source
|
* from which this server can be initialized.
|
*
|
* @param targetString The string representing the source
|
* @return The source as a integer value
|
* @throws DirectoryException if the string is not valid
|
*/
|
public int decodeTarget(String targetString) throws DirectoryException
|
{
|
if (targetString.equalsIgnoreCase("all"))
|
{
|
return RoutableMsg.ALL_SERVERS;
|
}
|
|
// So should be a serverID
|
try
|
{
|
int target = Integer.decode(targetString);
|
if (target >= 0)
|
{
|
// FIXME Could we check now that it is a know server in the domain ?
|
}
|
return target;
|
}
|
catch (Exception e)
|
{
|
ResultCode resultCode = ResultCode.OTHER;
|
Message message = ERR_INVALID_EXPORT_TARGET.get();
|
throw new DirectoryException(resultCode, message, e);
|
}
|
}
|
|
/**
|
* Initializes a remote server from this server.
|
* <p>
|
* The {@code exportBackend(OutputStream)} will therefore be called
|
* on this server, and the {@code importBackend(InputStream)}
|
* will be called on the remote server.
|
* <p>
|
* The InputStream and OutpuStream given as a parameter to those
|
* methods will be connected through the replication protocol.
|
*
|
* @param target The server-id of the server that should be initialized.
|
* The target can be discovered using the
|
* {@link #getReplicasList()} method.
|
* @param initTask The task that triggers this initialization and that should
|
* be updated with its progress.
|
*
|
* @throws DirectoryException If it was not possible to publish the
|
* Initialization message to the Topology.
|
*/
|
public void initializeRemote(int target, Task initTask)
|
throws DirectoryException
|
{
|
|
initializeRemote(target, this.serverID, initTask, this.initWindow);
|
|
|
}
|
|
/**
|
* Process the initialization of some other server or servers in the topology
|
* specified by the target argument when this initialization specifying the
|
* server that requests the initialization.
|
*
|
* @param serverToInitialize The target server that should be initialized.
|
* @param serverRunningTheTask The server that initiated the export. It can
|
* be the serverID of this server, or the serverID of a remote server.
|
* @param initTask The task in this server that triggers this initialization
|
* and that should be updated with its progress. Null when the export is done
|
* following a request coming from a remote server (task is remote).
|
* @param initWindow The value of the initialization window for flow control
|
* between the importer and the exporter.
|
*
|
* @exception DirectoryException When an error occurs. No exception raised
|
* means success.
|
*/
|
protected void initializeRemote(int serverToInitialize,
|
int serverRunningTheTask, Task initTask, int initWindow)
|
throws DirectoryException
|
{
|
DirectoryException exportRootException = null;
|
boolean contextAcquired = false;
|
|
// Acquire and initialize the export context
|
acquireIEContext(false);
|
contextAcquired = true;
|
|
// We manage the list of servers to initialize in order :
|
// - to test at the end that all expected servers have reconnected
|
// after their import and with the right genId
|
// - to update the task with the server(s) where this test failed
|
|
if (serverToInitialize == RoutableMsg.ALL_SERVERS)
|
{
|
Message msg = NOTE_FULL_UPDATE_ENGAGED_FOR_REMOTE_START_ALL.get(
|
countEntries(), serviceID, serverID);
|
logError(msg);
|
|
for (DSInfo dsi : getReplicasList())
|
ieContext.startList.add(dsi.getDsId());
|
|
// We manage the list of servers with which a flow control can be enabled
|
for (DSInfo dsi : getReplicasList())
|
{
|
if (dsi.getProtocolVersion()>= ProtocolVersion.REPLICATION_PROTOCOL_V4)
|
ieContext.setAckVal(dsi.getDsId(), 0);
|
}
|
}
|
else
|
{
|
Message msg = NOTE_FULL_UPDATE_ENGAGED_FOR_REMOTE_START.get(
|
countEntries(), serviceID, serverID, serverToInitialize);
|
logError(msg);
|
|
ieContext.startList.add(serverToInitialize);
|
|
// We manage the list of servers with which a flow control can be enabled
|
for (DSInfo dsi : getReplicasList())
|
{
|
if (dsi.getDsId() == serverToInitialize)
|
if (dsi.getProtocolVersion()>= ProtocolVersion.REPLICATION_PROTOCOL_V4)
|
ieContext.setAckVal(dsi.getDsId(), 0);
|
}
|
}
|
|
// loop for the case where the exporter is the initiator
|
int attempt = 0;
|
boolean done = false;
|
while ((!done) && (++attempt<2)) // attempt loop
|
{
|
try
|
{
|
ieContext.exportTarget = serverToInitialize;
|
if (initTask != null)
|
ieContext.initializeTask = initTask;
|
ieContext.initializeCounters(this.countEntries());
|
ieContext.msgCnt = 0;
|
ieContext.initNumLostConnections = broker.getNumLostConnections();
|
ieContext.initWindow = initWindow;
|
|
// Send start message to the peer
|
InitializeTargetMsg initTargetMsg = new InitializeTargetMsg(
|
serviceID, serverID, serverToInitialize, serverRunningTheTask,
|
ieContext.entryCount, initWindow);
|
|
broker.publish(initTargetMsg);
|
|
// Wait for all servers to be ok
|
waitForRemoteStartOfInit();
|
|
// Servers that left in the list are those for which we could not test
|
// that they have been successfully initialized.
|
if (!ieContext.failureList.isEmpty())
|
{
|
throw new DirectoryException(
|
ResultCode.OTHER,
|
ERR_INIT_NO_SUCCESS_START_FROM_SERVERS.get(
|
ieContext.failureList.toString()));
|
}
|
|
exportBackend(new BufferedOutputStream(new ReplOutputStream(this)));
|
|
// Notify the peer of the success
|
DoneMsg doneMsg = new DoneMsg(serverID, initTargetMsg.getDestination());
|
broker.publish(doneMsg);
|
|
}
|
catch(DirectoryException exportException)
|
{
|
// Give priority to the first exception raised - stored in the context
|
if (ieContext.exception != null)
|
exportRootException = ieContext.exception;
|
else
|
exportRootException = exportException;
|
}
|
|
if (debugEnabled())
|
TRACER.debugInfo(
|
"[IE] In " + this.monitor.getMonitorInstanceName()
|
+ " export ends with "
|
+ " connected=" + broker.isConnected()
|
+ " exportRootException=" + exportRootException);
|
|
if (exportRootException != null)
|
{
|
try
|
{
|
// Handling the errors during export
|
|
// Note: we could have lost the connection and another thread
|
// the listener one) has already managed to reconnect.
|
// So we MUST rely on the test broker.isConnected()
|
// ONLY to do 'wait to be reconnected by another thread'
|
// (if not yet reconnected already).
|
|
if (!broker.isConnected())
|
{
|
// We are still disconnected, so we wait for the listener thread
|
// to reconnect - wait 10s
|
if (debugEnabled())
|
TRACER.debugInfo(
|
"[IE] Exporter wait for reconnection by the listener thread");
|
int att=0;
|
while ((!broker.shuttingDown()) &&
|
(!broker.isConnected())&& (++att<100))
|
try { Thread.sleep(100); } catch(Exception e){}
|
}
|
|
if ((initTask != null) && broker.isConnected() &&
|
(serverToInitialize != RoutableMsg.ALL_SERVERS))
|
{
|
// NewAttempt case : In the case where
|
// - it's not an InitializeAll
|
// - AND the previous export attempt failed
|
// - AND we are (now) connected
|
// - and we own the task and this task is not an InitializeAll
|
// Let's :
|
// - sleep to let time to the other peer to reconnect if needed
|
// - and launch another attempt
|
try { Thread.sleep(1000); } catch(Exception e){}
|
logError(NOTE_RESENDING_INIT_TARGET.get((exportRootException!=null?
|
exportRootException.getLocalizedMessage():"")));
|
|
continue;
|
}
|
|
ErrorMsg errorMsg =
|
new ErrorMsg(serverToInitialize,
|
exportRootException.getMessageObject());
|
broker.publish(errorMsg);
|
}
|
catch(Exception e)
|
{
|
// Ignore the failure raised while proceeding the root failure
|
}
|
}
|
|
// We are always done for this export ...
|
// ... except in the NewAttempt case (see above)
|
done = true;
|
|
} // attempt loop
|
|
// Wait for all servers to be ok, and build the failure list
|
waitForRemoteEndOfInit();
|
|
// Servers that left in the list are those for which we could not test
|
// that they have been successfully initialized.
|
if (!ieContext.failureList.isEmpty())
|
{
|
if (exportRootException == null)
|
exportRootException = new DirectoryException(ResultCode.OTHER,
|
ERR_INIT_NO_SUCCESS_END_FROM_SERVERS.get(
|
Long.toString(getGenerationID()),
|
ieContext.failureList.toString()));
|
}
|
|
if (contextAcquired)
|
releaseIEContext();
|
|
String cause = exportRootException != null ? exportRootException
|
.getLocalizedMessage() : "";
|
if (serverToInitialize == RoutableMsg.ALL_SERVERS)
|
{
|
Message msg = NOTE_FULL_UPDATE_ENGAGED_FOR_REMOTE_END_ALL
|
.get(serviceID, serverID, cause);
|
logError(msg);
|
}
|
else
|
{
|
Message msg = NOTE_FULL_UPDATE_ENGAGED_FOR_REMOTE_END.get(
|
serviceID, serverID, serverToInitialize, cause);
|
logError(msg);
|
}
|
|
|
if (exportRootException != null)
|
{
|
throw(exportRootException);
|
}
|
|
}
|
|
/*
|
* For all remote servers in tht start list,
|
* - wait it has finished the import and present the expected generationID
|
* - build the failureList
|
*/
|
private void waitForRemoteStartOfInit()
|
{
|
int waitResultAttempt = 0;
|
Set<Integer> replicasWeAreWaitingFor = new HashSet<Integer>(0);
|
|
for (Integer sid : ieContext.startList)
|
replicasWeAreWaitingFor.add(sid);
|
|
if (debugEnabled())
|
TRACER.debugInfo(
|
"[IE] wait for start replicasWeAreWaitingFor=" + replicasWeAreWaitingFor);
|
|
boolean done = true;
|
do
|
{
|
done = true;
|
for (DSInfo dsi : getReplicasList())
|
{
|
if (debugEnabled())
|
TRACER.debugInfo(
|
"[IE] wait for start dsid " + dsi.getDsId()
|
+ " " + dsi.getStatus()
|
+ " " + dsi.getGenerationId()
|
+ " " + this.getGenerationID());
|
if (ieContext.startList.contains(dsi.getDsId()))
|
{
|
if (dsi.getStatus() != ServerStatus.FULL_UPDATE_STATUS)
|
{
|
// this one is still not doing the Full Update ... retry later
|
done = false;
|
try
|
{ Thread.sleep(100); } catch (InterruptedException e) {}
|
waitResultAttempt++;
|
break;
|
}
|
else
|
{
|
// this one is ok
|
replicasWeAreWaitingFor.remove(dsi.getDsId());
|
}
|
}
|
}
|
}
|
while ((!done) && (waitResultAttempt<1200) // 2mn
|
&& (!broker.shuttingDown()));
|
|
// Add to the failure list the servers that were here at start time but
|
// that never ended with the right generationId.
|
for (Integer sid : replicasWeAreWaitingFor.toArray(new Integer[0]))
|
ieContext.failureList.add(sid);
|
|
if (debugEnabled())
|
TRACER.debugInfo(
|
"[IE] wait for start ends with " + ieContext.failureList);
|
}
|
|
/*
|
* For all remote servers in the start list,
|
* - wait it has finished the import and present the expected generationID
|
* - build the failureList
|
*/
|
private void waitForRemoteEndOfInit()
|
{
|
Set<Integer> replicasWeAreWaitingFor = new HashSet<Integer>(0);
|
|
for (Integer sid : ieContext.startList)
|
replicasWeAreWaitingFor.add(sid);
|
|
if (debugEnabled())
|
TRACER.debugInfo(
|
"[IE] wait for end replicasWeAreWaitingFor=" + replicasWeAreWaitingFor);
|
|
// In case some new servers appear during the init, we want them to be
|
// considered in the processing of sorting the successfully initialized
|
// and the others
|
for (DSInfo dsi : getReplicasList())
|
replicasWeAreWaitingFor.add(dsi.getDsId());
|
|
boolean done = true;
|
do
|
{
|
done = true;
|
short reconnectMaxDelayInSec = 10;
|
short reconnectWait = 0;
|
Integer[] servers = replicasWeAreWaitingFor.toArray(new Integer[0]);
|
for (int serverId : servers)
|
{
|
if (ieContext.failureList.contains(serverId))
|
{
|
// this server has already been in error during initialization
|
// dont't wait for it
|
continue;
|
}
|
|
DSInfo dsInfo = null;
|
dsInfo = isRemoteDSConnected(serverId);
|
if (dsInfo == null)
|
{
|
// this server is disconnected
|
// may be for a long time if it crashed or had been stopped
|
// may be just the time to reconnect after import : should be short
|
if (++reconnectWait<reconnectMaxDelayInSec)
|
{
|
// let's still wait to give a chance to this server to reconnect
|
done = false;
|
}
|
else
|
{
|
// we left enough time to the servers to reconnect - now it's too
|
// late
|
}
|
}
|
else
|
{
|
// this server is connected
|
if (dsInfo.getStatus() == ServerStatus.FULL_UPDATE_STATUS)
|
{
|
// this one is still doing the Full Update ... retry later
|
done = false;
|
break;
|
}
|
else
|
{
|
// this one is done with the Full Update
|
if (dsInfo.getGenerationId() == this.getGenerationID())
|
{
|
// and with the expected generationId
|
replicasWeAreWaitingFor.remove(serverId);
|
}
|
}
|
}
|
}
|
|
// loop and wait
|
if (!done)
|
try { Thread.sleep(1000); } catch (InterruptedException e) {} // 1sec
|
|
}
|
while ((!done) && (!broker.shuttingDown())); // infinite wait
|
|
// Add to the failure list the servers that were here at start time but
|
// that never ended with the right generationId.
|
for (Integer sid : replicasWeAreWaitingFor.toArray(new Integer[0]))
|
ieContext.failureList.add(sid);
|
|
if (debugEnabled())
|
TRACER.debugInfo(
|
"[IE] wait for end ends with " + ieContext.failureList);
|
|
|
}
|
|
/**
|
* Get the ServerState maintained by the Concrete class.
|
*
|
* @return the ServerState maintained by the Concrete class.
|
*/
|
public ServerState getServerState()
|
{
|
return state;
|
}
|
|
|
private synchronized void acquireIEContext(boolean importInProgress)
|
throws DirectoryException
|
{
|
if (ieContext != null)
|
{
|
// Rejects 2 simultaneous exports
|
Message message = ERR_SIMULTANEOUS_IMPORT_EXPORT_REJECTED.get();
|
throw new DirectoryException(ResultCode.OTHER,
|
message);
|
}
|
|
ieContext = new IEContext(importInProgress);
|
}
|
|
private synchronized void releaseIEContext()
|
{
|
ieContext = null;
|
}
|
|
/**
|
* Processes an error message received while an export is
|
* on going, or an import will start.
|
*
|
* @param errorMsg The error message received.
|
*/
|
private void processErrorMsg(ErrorMsg errorMsg)
|
{
|
if (ieContext != null)
|
{
|
if (ieContext.exportTarget != RoutableMsg.ALL_SERVERS)
|
{
|
// The ErrorMsg is received while we have started an initialization
|
if (ieContext.getException() == null)
|
ieContext.setException(new DirectoryException(ResultCode.OTHER,
|
errorMsg.getDetails()));
|
|
/*
|
* This can happen :
|
* - on the first InitReqMsg sent when source in not known for example
|
* - on the next attempt when source crashed and did not reconnect
|
* even after the nextInitAttemptDelay
|
* During the import, the ErrorMsg will be received by receiveEntryBytes
|
*/
|
if (ieContext.initializeTask instanceof InitializeTask)
|
{
|
// Update the task that initiated the import
|
((InitializeTask)ieContext.initializeTask).
|
updateTaskCompletionState(ieContext.getException());
|
|
releaseIEContext();
|
}
|
}
|
else
|
{
|
// When we are the exporter in the case of initializeAll
|
// exporting must not be stopped on the first error.
|
}
|
}
|
}
|
|
/**
|
* Receives bytes related to an entry in the context of an import to
|
* initialize the domain (called by ReplLDIFInputStream).
|
*
|
* @return The bytes. Null when the Done or Err message has been received
|
*/
|
protected byte[] receiveEntryBytes()
|
{
|
ReplicationMsg msg;
|
while (true)
|
{
|
try
|
{
|
// In the context of the total update, we don't want any automatic
|
// re-connection done transparently by the broker because of a better
|
// RS or because of a connection failure.
|
// We want to be notified of topology change in order to track a
|
// potential disconnection of the exporter.
|
msg = broker.receive(false, false, true);
|
|
if (debugEnabled())
|
TRACER.debugInfo(
|
"[IE] In " + this.monitor.getMonitorInstanceName() +
|
", receiveEntryBytes " + msg);
|
|
if (msg == null)
|
{
|
if (broker.shuttingDown())
|
{
|
// The server is in the shutdown process
|
return null;
|
}
|
else
|
{
|
// Handle connection issues
|
if (ieContext.getException() == null)
|
ieContext.setException(new DirectoryException(
|
ResultCode.OTHER,
|
ERR_INIT_RS_DISCONNECTION_DURING_IMPORT.get(
|
broker.getReplicationServer())));
|
return null;
|
}
|
}
|
|
// Check good sequentiality of msg received
|
if (msg instanceof EntryMsg)
|
{
|
EntryMsg entryMsg = (EntryMsg)msg;
|
byte[] entryBytes = entryMsg.getEntryBytes();
|
ieContext.updateCounters(countEntryLimits(entryBytes));
|
|
if (ieContext.exporterProtocolVersion >=
|
ProtocolVersion.REPLICATION_PROTOCOL_V4)
|
{
|
// check the msgCnt of the msg received to check sequenciality
|
if (++ieContext.msgCnt != entryMsg.getMsgId())
|
{
|
if (ieContext.getException() == null)
|
ieContext.setException(new DirectoryException(ResultCode.OTHER,
|
ERR_INIT_BAD_MSG_ID_SEQ_DURING_IMPORT.get(
|
String.valueOf(ieContext.msgCnt),
|
String.valueOf(entryMsg.getMsgId()))));
|
return null;
|
}
|
|
// send the ack of flow control mgmt
|
if ((ieContext.msgCnt % (ieContext.initWindow/2)) == 0)
|
{
|
InitializeRcvAckMsg amsg = new InitializeRcvAckMsg(
|
this.serverID,
|
entryMsg.getSenderID(),
|
ieContext.msgCnt);
|
broker.publish(amsg, false);
|
if (debugEnabled())
|
TRACER.debugInfo(
|
"[IE] In " + this.monitor.getMonitorInstanceName() +
|
", publish InitializeRcvAckMsg" + amsg);
|
}
|
}
|
return entryBytes;
|
}
|
else if (msg instanceof DoneMsg)
|
{
|
// This is the normal termination of the import
|
// No error is stored and the import is ended
|
// by returning null
|
return null;
|
}
|
else if (msg instanceof ErrorMsg)
|
{
|
// This is an error termination during the import
|
// The error is stored and the import is ended
|
// by returning null
|
if (ieContext.getException() == null)
|
{
|
ErrorMsg errMsg = (ErrorMsg)msg;
|
if (errMsg.getCreationTime() > ieContext.startTime)
|
{
|
ieContext.setException(
|
new DirectoryException(ResultCode.OTHER,errMsg.getDetails()));
|
return null;
|
}
|
}
|
}
|
else
|
{
|
// Other messages received during an import are trashed except
|
// the topologyMsg.
|
if ((msg instanceof TopologyMsg) &&
|
(isRemoteDSConnected(ieContext.importSource)==null))
|
{
|
Message errMsg =
|
Message.raw(Category.SYNC, Severity.NOTICE,
|
ERR_INIT_EXPORTER_DISCONNECTION.get(
|
this.serviceID,
|
Integer.toString(this.serverID),
|
Integer.toString(ieContext.importSource)));
|
if (ieContext.getException()==null)
|
ieContext.setException(new DirectoryException(ResultCode.OTHER,
|
errMsg));
|
return null;
|
}
|
}
|
}
|
catch(Exception e)
|
{
|
// TODO: i18n
|
if (ieContext.getException() == null)
|
ieContext.setException(new DirectoryException(ResultCode.OTHER,
|
ERR_INIT_IMPORT_FAILURE.get(e.getLocalizedMessage())));
|
}
|
}
|
}
|
|
/**
|
* Count the number of entries in the provided byte[].
|
* This is based on the hypothesis that the entries are separated
|
* by a "\n\n" String.
|
*
|
* @param entryBytes
|
* @return The number of entries in the provided byte[].
|
*/
|
private int countEntryLimits(byte[] entryBytes)
|
{
|
return countEntryLimits(entryBytes, 0, entryBytes.length);
|
}
|
|
/**
|
* Count the number of entries in the provided byte[].
|
* This is based on the hypothesis that the entries are separated
|
* by a "\n\n" String.
|
*
|
* @param entryBytes
|
* @return The number of entries in the provided byte[].
|
*/
|
private int countEntryLimits(byte[] entryBytes, int pos, int length)
|
{
|
int entryCount = 0;
|
int count = 0;
|
while (count<=length-2)
|
{
|
if ((entryBytes[pos+count] == '\n') && (entryBytes[pos+count+1] == '\n'))
|
{
|
entryCount++;
|
count++;
|
}
|
count++;
|
}
|
return entryCount;
|
}
|
|
/**
|
* Exports an entry in LDIF format.
|
*
|
* @param lDIFEntry The entry to be exported in byte[] form.
|
* @param pos The starting Position in the array.
|
* @param length Number of array elements to be copied.
|
*
|
* @throws IOException when an error occurred.
|
*/
|
public void exportLDIFEntry(byte[] lDIFEntry, int pos, int length)
|
throws IOException
|
{
|
if (debugEnabled())
|
TRACER.debugInfo("[IE] Entering exportLDIFEntry entry=" + lDIFEntry);
|
|
// build the message
|
EntryMsg entryMessage = new EntryMsg(
|
serverID,ieContext.getExportTarget(), lDIFEntry, pos, length,
|
++ieContext.msgCnt);
|
|
// Waiting the slowest loop
|
while (!broker.shuttingDown())
|
{
|
// If an error was raised - like receiving an ErrorMsg from a remote
|
// server that have been stored by the listener thread in the ieContext,
|
// we just abandon the export by throwing an exception.
|
if (ieContext.getException() != null)
|
throw(new IOException(ieContext.getException().getMessage()));
|
|
int slowestServerId = ieContext.getSlowestServer();
|
if (isRemoteDSConnected(slowestServerId)==null)
|
{
|
ieContext.setException(new DirectoryException(ResultCode.OTHER,
|
ERR_INIT_HEARTBEAT_LOST_DURING_EXPORT.get(
|
Integer.toString(ieContext.getSlowestServer()))));
|
// .. and abandon the export by throwing an exception.
|
IOException ioe =
|
new IOException("IOException with nested DirectoryException");
|
ioe.initCause(ieContext.getException());
|
throw ioe;
|
}
|
|
int ourLastExportedCnt = ieContext.msgCnt;
|
int slowestCnt = ieContext.ackVals.get(slowestServerId);
|
|
if (debugEnabled())
|
TRACER.debugInfo("[IE] Entering exportLDIFEntry waiting " +
|
" our=" + ourLastExportedCnt + " slowest=" + slowestCnt);
|
|
if ((ourLastExportedCnt - slowestCnt) > ieContext.initWindow)
|
{
|
if (debugEnabled())
|
TRACER.debugInfo("[IE] Entering exportLDIFEntry waiting");
|
|
// our export is too far beyond the slowest importer - let's wait
|
try { Thread.sleep(100); } catch(Exception e) {}
|
|
// process any connection error
|
if ((broker.hasConnectionError())||
|
(broker.getNumLostConnections()!= ieContext.initNumLostConnections))
|
{
|
// publish failed - store the error in the ieContext ...
|
DirectoryException de = new DirectoryException(ResultCode.OTHER,
|
ERR_INIT_RS_DISCONNECTION_DURING_EXPORT.get(
|
Integer.toString(broker.getRsServerId())));
|
if (ieContext.getException() == null)
|
ieContext.setException(de);
|
// .. and abandon the export by throwing an exception.
|
throw new IOException(de.getMessage());
|
}
|
}
|
else
|
{
|
if (debugEnabled())
|
TRACER.debugInfo("[IE] slowest got to us => stop waiting");
|
break;
|
}
|
} // Waiting the slowest loop
|
|
if (debugEnabled())
|
TRACER.debugInfo("[IE] Entering exportLDIFEntry pub entry=" + lDIFEntry);
|
|
// publish the message
|
boolean sent = broker.publish(entryMessage, false);
|
|
// process any publish error
|
if (((!sent)||
|
(broker.hasConnectionError()))||
|
(broker.getNumLostConnections() != ieContext.initNumLostConnections))
|
{
|
// publish failed - store the error in the ieContext ...
|
DirectoryException de = new DirectoryException(ResultCode.OTHER,
|
ERR_INIT_RS_DISCONNECTION_DURING_EXPORT.get(
|
Integer.toString(broker.getRsServerId())));
|
if (ieContext.getException() == null)
|
ieContext.setException(de);
|
// .. and abandon the export by throwing an exception.
|
throw new IOException(de.getMessage());
|
}
|
|
// publish succeeded
|
try
|
{
|
ieContext.updateCounters(countEntryLimits(lDIFEntry, pos, length));
|
}
|
catch (DirectoryException de)
|
{
|
// store the error in the ieContext ...
|
if (ieContext.getException() == null)
|
ieContext.setException(de);
|
// .. and abandon the export by throwing an exception.
|
throw new IOException(de.getMessage());
|
}
|
}
|
|
/**
|
* Initializes this domain from another source server.
|
* <p>
|
* When this method is called, a request for initialization will
|
* be sent to the source server asking for initialization.
|
* <p>
|
* The {@code exportBackend(OutputStream)} will therefore be called
|
* on the source server, and the {@code importBackend(InputStream)}
|
* will be called on his server.
|
* <p>
|
* The InputStream and OutpuStream given as a parameter to those
|
* methods will be connected through the replication protocol.
|
*
|
* @param source The server-id of the source from which to initialize.
|
* The source can be discovered using the
|
* {@link #getReplicasList()} method.
|
*
|
* @throws DirectoryException If it was not possible to publish the
|
* Initialization message to the Topology.
|
*/
|
public void initializeFromRemote(int source)
|
throws DirectoryException
|
{
|
initializeFromRemote(source, null);
|
}
|
|
/**
|
* Initializes a remote server from this server.
|
* <p>
|
* The {@code exportBackend(OutputStream)} will therefore be called
|
* on this server, and the {@code importBackend(InputStream)}
|
* will be called on the remote server.
|
* <p>
|
* The InputStream and OutpuStream given as a parameter to those
|
* methods will be connected through the replication protocol.
|
*
|
* @param target The server-id of the server that should be initialized.
|
* The target can be discovered using the
|
* {@link #getReplicasList()} method.
|
*
|
* @throws DirectoryException If it was not possible to publish the
|
* Initialization message to the Topology.
|
*/
|
public void initializeRemote(int target) throws DirectoryException
|
{
|
initializeRemote(target, null);
|
}
|
|
/**
|
* Initializes asynchronously this domain from a remote source server.
|
* Before returning from this call, for the provided task :
|
* - the progressing counters are updated during the initialization using
|
* setTotal() and setLeft().
|
* - the end of the initialization using updateTaskCompletionState().
|
* <p>
|
* When this method is called, a request for initialization is sent to the
|
* remote source server requesting initialization.
|
* <p>
|
*
|
* @param source The server-id of the source from which to initialize.
|
* The source can be discovered using the
|
* {@link #getReplicasList()} method.
|
*
|
* @param initTask The task that launched the initialization
|
* and should be updated of its progress.
|
*
|
* @throws DirectoryException If it was not possible to publish the
|
* Initialization message to the Topology.
|
* The task state is updated.
|
*/
|
public void initializeFromRemote(int source, Task initTask)
|
throws DirectoryException
|
{
|
Message errMsg = null;
|
|
if (debugEnabled())
|
TRACER.debugInfo("[IE] Entering initializeFromRemote for " + this);
|
|
if (!broker.isConnected())
|
{
|
errMsg = ERR_INITIALIZATION_FAILED_NOCONN.get(getServiceID());
|
}
|
|
// We must not test here whether the remote source is connected to
|
// the topology by testing if it stands in the replicas list since.
|
// In the case of a re-attempt of initialization, the listener thread is
|
// running this method directly coming from initailize() method and did
|
// not processed any topology message in between the failure and the
|
// new attempt.
|
try
|
{
|
// We must immediatly acquire a context to store the task inside
|
// The context will be used when we (the listener thread) will receive
|
// the InitializeTargetMsg, process the import, and at the end
|
// update the task.
|
|
acquireIEContext(true); //test and set if no import already in progress
|
ieContext.initializeTask = initTask;
|
ieContext.attemptCnt = 0;
|
ieContext.initReqMsgSent = new InitializeRequestMsg(
|
serviceID, serverID, source, this.initWindow);
|
|
// Publish Init request msg
|
broker.publish(ieContext.initReqMsgSent);
|
|
// The normal success processing is now to receive InitTargetMsg then
|
// entries from the remote server.
|
// The error cases are :
|
// - either local error immediatly caught below
|
// - a remote error we will receive as an ErrorMsg
|
}
|
catch(DirectoryException de)
|
{
|
errMsg = de.getMessageObject();
|
}
|
catch(Exception e)
|
{
|
// Should not happen
|
errMsg = Message.raw(Category.SYNC, Severity.NOTICE,
|
e.getLocalizedMessage());
|
logError(errMsg);
|
}
|
|
// When error, update the task and raise the error to the caller
|
if (errMsg != null)
|
{
|
// No need to call here updateTaskCompletionState - will be done
|
// by the caller
|
releaseIEContext();
|
DirectoryException de = new DirectoryException(
|
ResultCode.OTHER,
|
errMsg);
|
throw (de);
|
}
|
}
|
|
/**
|
* Processes an InitializeTargetMsg received from a remote server
|
* meaning processes an initialization from the entries expected to be
|
* received now.
|
*
|
* @param initTargetMsgReceived The message received from the remote server.
|
*
|
* @param requestorServerId The serverId of the server that requested the
|
* initialization meaning the server where the
|
* task has initially been created (this server,
|
* or the remote server).
|
*/
|
void initialize(InitializeTargetMsg initTargetMsgReceived,
|
int requestorServerId)
|
{
|
InitializeTask initFromtask = null;
|
|
if (debugEnabled())
|
TRACER.debugInfo("[IE] Entering initialize - domain=" + this);
|
|
int source = initTargetMsgReceived.getSenderID();
|
|
try
|
{
|
// Log starting
|
Message msg = NOTE_FULL_UPDATE_ENGAGED_FROM_REMOTE_START.get(
|
serviceID, initTargetMsgReceived.getSenderID(), serverID);
|
logError(msg);
|
|
// Go into full update status
|
setNewStatus(StatusMachineEvent.TO_FULL_UPDATE_STATUS_EVENT);
|
|
// Acquire an import context if no already done (and initialize).
|
if (initTargetMsgReceived.getInitiatorID() == this.serverID)
|
{
|
// The initTargetMsgReceived received is the answer to a request that
|
// we (this server) sent previously. In this case, so the IEContext
|
// has been already acquired when the request was published in order
|
// to store the task (to be updated with the status at the end).
|
}
|
else
|
{
|
// The initTargetMsgReceived is for an import initiated by the remote
|
// server.
|
// Test and set if no import already in progress
|
acquireIEContext(true);
|
}
|
|
// Initialize stuff
|
ieContext.importSource = source;
|
ieContext.initializeCounters(initTargetMsgReceived.getEntryCount());
|
ieContext.initWindow = initTargetMsgReceived.getInitWindow();
|
// Protocol version is -1 when not known.
|
ieContext.exporterProtocolVersion = getProtocolVersion(source);
|
initFromtask = (InitializeTask)ieContext.initializeTask;
|
|
// Lauch the import
|
importBackend(new ReplInputStream(this));
|
|
}
|
catch (DirectoryException e)
|
{
|
// Store the exception raised. It will be considered if no other exception
|
// has been previously stored in the context
|
if (ieContext.getException() == null)
|
ieContext.setException(e);
|
}
|
finally
|
{
|
if (debugEnabled())
|
TRACER.debugInfo("[IE] Domain=" + this
|
+ " ends import with exception=" + ieContext.getException()
|
+ " connected=" + broker.isConnected());
|
|
// It is necessary to restart (reconnect to RS) for different reasons
|
// - when everything went well, reconnect in order to exchange
|
// new state, new generation ID
|
// - when we have connection failure, reconnect to retry a new import
|
// right here, right now
|
// we never want retryOnFailure if we fails reconnecting in the restart.
|
broker.reStart(false);
|
|
if (ieContext.getException() != null)
|
{
|
if (broker.isConnected() && (initFromtask != null)
|
&& (++ieContext.attemptCnt<2))
|
{
|
// Worth a new attempt
|
// since initFromtask is in this server, connection is ok
|
try
|
{
|
|
// Wait for the exporter to stabilize - eventually reconnect as
|
// well if it was connected to the same RS than the one we lost ...
|
Thread.sleep(1000);
|
|
// Restart the whole import protocol exchange by sending again
|
// the request
|
logError(NOTE_RESENDING_INIT_FROM_REMOTE_REQUEST.get(
|
ieContext.getException().getLocalizedMessage()));
|
|
broker.publish(ieContext.initReqMsgSent);
|
|
ieContext.initializeCounters(0);
|
ieContext.exception = null;
|
ieContext.msgCnt = 0;
|
|
// Processing of the received initTargetMsgReceived is done
|
// let's wait for the next one
|
return;
|
}
|
catch(Exception e)
|
{
|
// An error occurs when sending a new request for a new import.
|
// This error is not stored, prefering to keep the initial one.
|
logError(ERR_SENDING_NEW_ATTEMPT_INIT_REQUEST.get(
|
e.getLocalizedMessage(),
|
ieContext.getException().getLocalizedMessage()));
|
}
|
}
|
}
|
|
// ===================
|
// No new attempt case
|
|
if (debugEnabled())
|
TRACER.debugInfo("[IE] Domain=" + this
|
+ " ends initialization with exception=" + ieContext.getException()
|
+ " connected=" + broker.isConnected()
|
+ " task=" + initFromtask
|
+ " attempt=" + ieContext.attemptCnt);
|
|
try
|
{
|
if (broker.isConnected() && (ieContext.getException() != null))
|
{
|
// Let's notify the exporter
|
ErrorMsg errorMsg = new ErrorMsg(requestorServerId,
|
ieContext.getException().getMessageObject());
|
broker.publish(errorMsg);
|
}
|
else // !broker.isConnected()
|
{
|
// Don't try to reconnect here.
|
// The current running thread is the listener thread and will loop on
|
// receive() that is expected to manage reconnects attempt.
|
}
|
|
// Update the task that initiated the import must be the last thing.
|
// Particularly, broker.restart() after import success must be done
|
// before some other operations/tasks to be launched,
|
// like resetting the generation ID.
|
if (initFromtask != null)
|
{
|
initFromtask.updateTaskCompletionState(ieContext.getException());
|
}
|
}
|
finally
|
{
|
Message msg = NOTE_FULL_UPDATE_ENGAGED_FROM_REMOTE_END.get(
|
serviceID, initTargetMsgReceived.getSenderID(), serverID,
|
(ieContext.getException() != null ? ieContext
|
.getException().getLocalizedMessage() : ""));
|
logError(msg);
|
releaseIEContext();
|
} // finally
|
} // finally
|
}
|
|
/**
|
* Return the protocol version of the DS related to the provided serverid.
|
* Returns -1 when the protocol version is not known.
|
* @param dsServerId The provided serverid.
|
* @return The procotol version.
|
*/
|
short getProtocolVersion(int dsServerId)
|
{
|
short protocolVersion = -1;
|
for (DSInfo dsi : getReplicasList())
|
{
|
if (dsi.getDsId() == dsServerId)
|
{
|
protocolVersion = dsi.getProtocolVersion();
|
break;
|
}
|
}
|
return protocolVersion;
|
}
|
|
/**
|
* Sets the status to a new value depending of the passed status machine
|
* event.
|
* @param event The event that may make the status be changed
|
*/
|
protected void setNewStatus(StatusMachineEvent event)
|
{
|
ServerStatus newStatus =
|
StatusMachine.computeNewStatus(status, event);
|
|
if (newStatus == ServerStatus.INVALID_STATUS)
|
{
|
Message msg = ERR_DS_CANNOT_CHANGE_STATUS.get(serviceID,
|
Integer.toString(serverID), status.toString(), event.toString());
|
logError(msg);
|
return;
|
}
|
|
if (newStatus != status)
|
{
|
// Reset status date
|
lastStatusChangeDate = new Date();
|
// Reset monitoring counters if reconnection
|
if (newStatus == ServerStatus.NOT_CONNECTED_STATUS)
|
resetMonitoringCounters();
|
|
// Store new status
|
status = newStatus;
|
|
if (debugEnabled())
|
TRACER.debugInfo("Replication domain " + serviceID +
|
" new status is: " + status);
|
|
// Perform whatever actions are needed to apply properties for being
|
// compliant with new status
|
updateDomainForNewStatus();
|
}
|
}
|
|
/**
|
* Returns a boolean indicating if an import or export is currently
|
* processed.
|
*
|
* @return The status
|
*/
|
public boolean ieRunning()
|
{
|
return (ieContext != null);
|
}
|
|
/**
|
* Check the value of the Replication Servers generation ID.
|
*
|
* @param generationID The expected value of the generation ID.
|
*
|
* @throws DirectoryException When the generation ID of the Replication
|
* Servers is not the expected value.
|
*/
|
private void checkGenerationID(long generationID)
|
throws DirectoryException
|
{
|
boolean allset = true;
|
|
for (int i = 0; i< 50; i++)
|
{
|
allset = true;
|
for (RSInfo rsInfo : getRsList())
|
{
|
// the 'empty' RSes (generationId==-1) are considered as good citizens
|
if ((rsInfo.getGenerationId() != -1) &&
|
(rsInfo.getGenerationId() != generationID))
|
{
|
try
|
{
|
Thread.sleep(i*100);
|
} catch (InterruptedException e)
|
{
|
}
|
allset = false;
|
break;
|
}
|
}
|
if (allset)
|
{
|
break;
|
}
|
}
|
if (!allset)
|
{
|
ResultCode resultCode = ResultCode.OTHER;
|
Message message = ERR_RESET_GENERATION_ID_FAILED.get(serviceID);
|
throw new DirectoryException(
|
resultCode, message);
|
}
|
}
|
|
/**
|
* Reset the Replication Log.
|
* Calling this method will remove all the Replication information that
|
* was kept on all the Replication Servers currently connected in the
|
* topology.
|
*
|
* @throws DirectoryException If this ReplicationDomain is not currently
|
* connected to a Replication Server or it
|
* was not possible to contact it.
|
*/
|
public void resetReplicationLog() throws DirectoryException
|
{
|
// Reset the Generation ID to -1 to clean the ReplicationServers.
|
resetGenerationId((long)-1);
|
|
// check that at least one ReplicationServer did change its generation-id
|
checkGenerationID(-1);
|
|
// Reconnect to the Replication Server so that it adopt our
|
// GenerationID.
|
disableService();
|
enableService();
|
|
// wait for the domain to reconnect.
|
int count = 0;
|
while (!isConnected() && (count < 10))
|
{
|
try
|
{
|
Thread.sleep(100);
|
} catch (InterruptedException e)
|
{
|
}
|
}
|
|
resetGenerationId(getGenerationID());
|
|
// check that at least one ReplicationServer did change its generation-id
|
checkGenerationID(getGenerationID());
|
}
|
|
/**
|
* Reset the generationId of this domain in the whole topology.
|
* A message is sent to the Replication Servers for them to reset
|
* their change dbs.
|
*
|
* @param generationIdNewValue The new value of the generation Id.
|
* @throws DirectoryException When an error occurs
|
*/
|
public void resetGenerationId(Long generationIdNewValue)
|
throws DirectoryException
|
{
|
if (debugEnabled())
|
TRACER.debugInfo(
|
"Server id " + serverID + " and domain " + serviceID
|
+ " resetGenerationId " + generationIdNewValue);
|
|
ResetGenerationIdMsg genIdMessage = null;
|
|
if (generationIdNewValue == null)
|
{
|
genIdMessage = new ResetGenerationIdMsg(this.getGenerationID());
|
}
|
else
|
{
|
genIdMessage = new ResetGenerationIdMsg(generationIdNewValue);
|
}
|
|
if (!isConnected())
|
{
|
ResultCode resultCode = ResultCode.OTHER;
|
Message message = ERR_RESET_GENERATION_CONN_ERR_ID.get(serviceID,
|
Integer.toString(serverID),
|
Long.toString(genIdMessage.getGenerationId()));
|
throw new DirectoryException(
|
resultCode, message);
|
}
|
broker.publish(genIdMessage);
|
|
// check that at least one ReplicationServer did change its generation-id
|
if (generationIdNewValue == null)
|
{
|
checkGenerationID(this.getGenerationID());
|
}
|
else
|
{
|
checkGenerationID(generationIdNewValue);
|
}
|
}
|
|
|
/*
|
******** End of The total Update code *********
|
*/
|
|
/*
|
******* Start of Monitoring Code **********
|
*/
|
|
/**
|
* Get the maximum receive window size.
|
*
|
* @return The maximum receive window size.
|
*/
|
int getMaxRcvWindow()
|
{
|
if (broker != null)
|
return broker.getMaxRcvWindow();
|
else
|
return 0;
|
}
|
|
/**
|
* Get the current receive window size.
|
*
|
* @return The current receive window size.
|
*/
|
int getCurrentRcvWindow()
|
{
|
if (broker != null)
|
return broker.getCurrentRcvWindow();
|
else
|
return 0;
|
}
|
|
/**
|
* Get the maximum send window size.
|
*
|
* @return The maximum send window size.
|
*/
|
int getMaxSendWindow()
|
{
|
if (broker != null)
|
return broker.getMaxSendWindow();
|
else
|
return 0;
|
}
|
|
/**
|
* Get the current send window size.
|
*
|
* @return The current send window size.
|
*/
|
int getCurrentSendWindow()
|
{
|
if (broker != null)
|
return broker.getCurrentSendWindow();
|
else
|
return 0;
|
}
|
|
/**
|
* Get the number of times the replication connection was lost.
|
* @return The number of times the replication connection was lost.
|
*/
|
int getNumLostConnections()
|
{
|
if (broker != null)
|
return broker.getNumLostConnections();
|
else
|
return 0;
|
}
|
|
/**
|
* Determine whether the connection to the replication server is encrypted.
|
* @return true if the connection is encrypted, false otherwise.
|
*/
|
boolean isSessionEncrypted()
|
{
|
if (broker != null)
|
return broker.isSessionEncrypted();
|
else
|
return false;
|
}
|
|
/**
|
* This method is called when the ReplicationDomain has completed the
|
* processing of a received update synchronously.
|
* In such cases the processUpdateDone () is called and the state
|
* is updated automatically.
|
*
|
* @param msg The UpdateMessage that was processed.
|
*/
|
void processUpdateDoneSynchronous(UpdateMsg msg)
|
{
|
// Warning: in synchronous mode, no way to tell the replay of an update went
|
// wrong Just put null in processUpdateDone so that if assured replication
|
// is used the ack is sent without error at replay flag.
|
processUpdateDone(msg, null);
|
state.update(msg.getChangeNumber());
|
}
|
|
/**
|
* Check if the domain is connected to a ReplicationServer.
|
*
|
* @return true if the server is connected, false if not.
|
*/
|
public boolean isConnected()
|
{
|
if (broker != null)
|
return broker.isConnected();
|
else
|
return false;
|
}
|
|
/**
|
* Check if the domain has a connection error.
|
* A Connection error happens when the broker could not be created
|
* or when the broker could not find any ReplicationServer to connect to.
|
*
|
* @return true if the domain has a connection error.
|
*/
|
public boolean hasConnectionError()
|
{
|
if (broker != null)
|
return broker.hasConnectionError();
|
else
|
return true;
|
}
|
|
/**
|
* Get the name of the replicationServer to which this domain is currently
|
* connected.
|
*
|
* @return the name of the replicationServer to which this domain
|
* is currently connected.
|
*/
|
public String getReplicationServer()
|
{
|
if (broker != null)
|
return broker.getReplicationServer();
|
else
|
return "Not connected";
|
}
|
|
/**
|
* Gets the number of updates sent in assured safe read mode.
|
* @return The number of updates sent in assured safe read mode.
|
*/
|
public int getAssuredSrSentUpdates()
|
{
|
return assuredSrSentUpdates.get();
|
}
|
|
/**
|
* Gets the number of updates sent in assured safe read mode that have been
|
* acknowledged without errors.
|
* @return The number of updates sent in assured safe read mode that have been
|
* acknowledged without errors.
|
*/
|
public int getAssuredSrAcknowledgedUpdates()
|
{
|
return assuredSrAcknowledgedUpdates.get();
|
}
|
|
/**
|
* Gets the number of updates sent in assured safe read mode that have not
|
* been acknowledged.
|
* @return The number of updates sent in assured safe read mode that have not
|
* been acknowledged.
|
*/
|
public int getAssuredSrNotAcknowledgedUpdates()
|
{
|
return assuredSrNotAcknowledgedUpdates.get();
|
}
|
|
/**
|
* Gets the number of updates sent in assured safe read mode that have not
|
* been acknowledged due to timeout error.
|
* @return The number of updates sent in assured safe read mode that have not
|
* been acknowledged due to timeout error.
|
*/
|
public int getAssuredSrTimeoutUpdates()
|
{
|
return assuredSrTimeoutUpdates.get();
|
}
|
|
/**
|
* Gets the number of updates sent in assured safe read mode that have not
|
* been acknowledged due to wrong status error.
|
* @return The number of updates sent in assured safe read mode that have not
|
* been acknowledged due to wrong status error.
|
*/
|
public int getAssuredSrWrongStatusUpdates()
|
{
|
return assuredSrWrongStatusUpdates.get();
|
}
|
|
/**
|
* Gets the number of updates sent in assured safe read mode that have not
|
* been acknowledged due to replay error.
|
* @return The number of updates sent in assured safe read mode that have not
|
* been acknowledged due to replay error.
|
*/
|
public int getAssuredSrReplayErrorUpdates()
|
{
|
return assuredSrReplayErrorUpdates.get();
|
}
|
|
/**
|
* Gets the number of updates sent in assured safe read mode that have not
|
* been acknowledged per server.
|
* @return The number of updates sent in assured safe read mode that have not
|
* been acknowledged per server.
|
*/
|
public Map<Integer, Integer> getAssuredSrServerNotAcknowledgedUpdates()
|
{
|
// Clone a snapshot with synchronized section to have a consistent view in
|
// monitoring
|
Map<Integer, Integer> snapshot = new HashMap<Integer, Integer>();
|
synchronized(assuredSrServerNotAcknowledgedUpdates)
|
{
|
Set<Integer> keySet = assuredSrServerNotAcknowledgedUpdates.keySet();
|
for (Integer serverId : keySet)
|
{
|
Integer i = assuredSrServerNotAcknowledgedUpdates.get(serverId);
|
snapshot.put(serverId, i);
|
}
|
}
|
return snapshot;
|
}
|
|
/**
|
* Gets the number of updates received in assured safe read mode request.
|
* @return The number of updates received in assured safe read mode request.
|
*/
|
public int getAssuredSrReceivedUpdates()
|
{
|
return assuredSrReceivedUpdates.get();
|
}
|
|
/**
|
* Gets the number of updates received in assured safe read mode that we acked
|
* without error (no replay error).
|
* @return The number of updates received in assured safe read mode that we
|
* acked without error (no replay error).
|
*/
|
public int getAssuredSrReceivedUpdatesAcked()
|
{
|
return this.assuredSrReceivedUpdatesAcked.get();
|
}
|
|
/**
|
* Gets the number of updates received in assured safe read mode that we did
|
* not ack due to error (replay error).
|
* @return The number of updates received in assured safe read mode that we
|
* did not ack due to error (replay error).
|
*/
|
public int getAssuredSrReceivedUpdatesNotAcked()
|
{
|
return this.assuredSrReceivedUpdatesNotAcked.get();
|
}
|
|
/**
|
* Gets the number of updates sent in assured safe data mode.
|
* @return The number of updates sent in assured safe data mode.
|
*/
|
public int getAssuredSdSentUpdates()
|
{
|
return assuredSdSentUpdates.get();
|
}
|
|
/**
|
* Gets the number of updates sent in assured safe data mode that have been
|
* acknowledged without errors.
|
* @return The number of updates sent in assured safe data mode that have been
|
* acknowledged without errors.
|
*/
|
public int getAssuredSdAcknowledgedUpdates()
|
{
|
return assuredSdAcknowledgedUpdates.get();
|
}
|
|
/**
|
* Gets the number of updates sent in assured safe data mode that have not
|
* been acknowledged due to timeout error.
|
* @return The number of updates sent in assured safe data mode that have not
|
* been acknowledged due to timeout error.
|
*/
|
public int getAssuredSdTimeoutUpdates()
|
{
|
return assuredSdTimeoutUpdates.get();
|
}
|
|
/**
|
* Gets the number of updates sent in assured safe data mode that have not
|
* been acknowledged due to timeout error per server.
|
* @return The number of updates sent in assured safe data mode that have not
|
* been acknowledged due to timeout error per server.
|
*/
|
public Map<Integer, Integer> getAssuredSdServerTimeoutUpdates()
|
{
|
// Clone a snapshot with synchronized section to have a consistent view in
|
// monitoring
|
Map<Integer, Integer> snapshot = new HashMap<Integer, Integer>();
|
synchronized(assuredSdServerTimeoutUpdates)
|
{
|
Set<Integer> keySet = assuredSdServerTimeoutUpdates.keySet();
|
for (Integer serverId : keySet)
|
{
|
Integer i = assuredSdServerTimeoutUpdates.get(serverId);
|
snapshot.put(serverId, i);
|
}
|
}
|
return snapshot;
|
}
|
|
/**
|
* Gets the date of the last status change.
|
* @return The date of the last status change.
|
*/
|
public Date getLastStatusChangeDate()
|
{
|
return lastStatusChangeDate;
|
}
|
|
/**
|
* Resets the values of the monitoring counters.
|
*/
|
private void resetMonitoringCounters()
|
{
|
numProcessedUpdates = new AtomicInteger(0);
|
numRcvdUpdates = new AtomicInteger(0);
|
numSentUpdates = new AtomicInteger(0);
|
|
assuredSrSentUpdates = new AtomicInteger(0);
|
assuredSrAcknowledgedUpdates = new AtomicInteger(0);
|
assuredSrNotAcknowledgedUpdates = new AtomicInteger(0);
|
assuredSrTimeoutUpdates = new AtomicInteger(0);
|
assuredSrWrongStatusUpdates = new AtomicInteger(0);
|
assuredSrReplayErrorUpdates = new AtomicInteger(0);
|
assuredSrServerNotAcknowledgedUpdates = new HashMap<Integer,Integer>();
|
assuredSrReceivedUpdates = new AtomicInteger(0);
|
assuredSrReceivedUpdatesAcked = new AtomicInteger(0);
|
assuredSrReceivedUpdatesNotAcked = new AtomicInteger(0);
|
assuredSdSentUpdates = new AtomicInteger(0);
|
assuredSdAcknowledgedUpdates = new AtomicInteger(0);
|
assuredSdTimeoutUpdates = new AtomicInteger(0);
|
assuredSdServerTimeoutUpdates = new HashMap<Integer,Integer>();
|
}
|
|
/*
|
********** End of Monitoring Code **************
|
*/
|
|
/**
|
* Start the publish mechanism of the Replication Service.
|
* After this method has been called, the publish service can be used
|
* by calling the {@link #publish(UpdateMsg)} method.
|
*
|
* @param replicationServers The replication servers that should be used.
|
* @param window The window size of this replication domain.
|
* @param heartbeatInterval The heartbeatInterval that should be used
|
* to check the availability of the replication
|
* servers.
|
* @param changetimeHeartbeatInterval The interval used to send change
|
* time heartbeat to the replication server.
|
*
|
* @throws ConfigException If the DirectoryServer configuration was
|
* incorrect.
|
*/
|
public void startPublishService(
|
Collection<String> replicationServers, int window,
|
long heartbeatInterval, long changetimeHeartbeatInterval)
|
throws ConfigException
|
{
|
synchronized (sessionLock)
|
{
|
if (broker == null)
|
{
|
/*
|
* create the broker object used to publish and receive changes
|
*/
|
broker = new ReplicationBroker(
|
this, state, serviceID,
|
serverID, window,
|
getGenerationID(),
|
heartbeatInterval,
|
new ReplSessionSecurity(),
|
getGroupId(),
|
changetimeHeartbeatInterval);
|
|
broker.start(replicationServers);
|
|
/*
|
* Create a replication monitor object responsible for publishing
|
* monitoring information below cn=monitor.
|
*/
|
monitor = new ReplicationMonitor(this);
|
}
|
|
DirectoryServer.registerMonitorProvider(monitor);
|
}
|
}
|
|
/**
|
* Starts the receiver side of the Replication Service.
|
* <p>
|
* After this method has been called, the Replication Service will start
|
* calling the {@link #processUpdate(UpdateMsg)}.
|
* <p>
|
* This method must be called once and must be called after the
|
* {@link #startPublishService(Collection, int, long, long)}.
|
*
|
*/
|
public void startListenService()
|
{
|
synchronized (sessionLock)
|
{
|
//
|
// Create the listener thread
|
listenerThread = new ListenerThread(this);
|
listenerThread.start();
|
}
|
}
|
|
/**
|
* Temporarily disable the Replication Service.
|
* The Replication Service can be enabled again using
|
* {@link #enableService()}.
|
* <p>
|
* It can be useful to disable the Replication Service when the
|
* repository where the replicated information is stored becomes
|
* temporarily unavailable and replicated updates can therefore not
|
* be replayed during a while. This method is not MT safe.
|
*/
|
public void disableService()
|
{
|
synchronized (sessionLock)
|
{
|
// Stop the listener thread
|
if (listenerThread != null)
|
{
|
listenerThread.shutdown();
|
}
|
|
if (broker != null)
|
{
|
broker.stop();
|
}
|
|
// Wait for the listener thread to stop
|
if (listenerThread != null)
|
listenerThread.waitForShutdown();
|
}
|
}
|
|
/**
|
* Restart the Replication service after a {@link #disableService()}.
|
* <p>
|
* The Replication Service will restart from the point indicated by the
|
* {@link ServerState} that was given as a parameter to the
|
* {@link #startPublishService(Collection, int, long, long)}
|
* at startup time.
|
* If some data have changed in the repository during the period of time when
|
* the Replication Service was disabled, this {@link ServerState} should
|
* therefore be updated by the Replication Domain subclass before calling
|
* this method. This method is not MT safe.
|
*/
|
public void enableService()
|
{
|
synchronized (sessionLock)
|
{
|
broker.start();
|
|
// Create the listener thread
|
listenerThread = new ListenerThread(this);
|
listenerThread.start();
|
}
|
}
|
|
/**
|
* Definitively stops the Replication Service.
|
*/
|
public void stopDomain()
|
{
|
DirectoryServer.deregisterMonitorProvider(monitor.getMonitorInstanceName());
|
|
disableService();
|
domains.remove(serviceID);
|
}
|
|
/**
|
* Change some ReplicationDomain parameters.
|
*
|
* @param replicationServers The new list of Replication Servers that this
|
* domain should now use.
|
* @param windowSize The window size that this domain should use.
|
* @param heartbeatInterval The heartbeatInterval that this domain should
|
* use.
|
* @param groupId The new group id to use
|
*/
|
public void changeConfig(
|
Collection<String> replicationServers,
|
int windowSize,
|
long heartbeatInterval,
|
byte groupId)
|
{
|
this.groupId = groupId;
|
|
if (broker != null)
|
{
|
if (broker.changeConfig(
|
replicationServers, windowSize, heartbeatInterval, groupId))
|
{
|
disableService();
|
enableService();
|
}
|
}
|
}
|
|
/**
|
* Change some ReplicationDomain parameters : the ECL include attribute.
|
*
|
* @param newECLInclude The new ECL attribute.
|
*/
|
public void changeConfig(Set<String> newECLInclude)
|
{
|
boolean configECLIncludeChanged = false;
|
Set<String> currentECLInclude = this.getEclInclude(serverID);
|
|
if (newECLInclude.size() != currentECLInclude.size())
|
{
|
configECLIncludeChanged = true;
|
}
|
else
|
{
|
// compare current config and new config
|
for (String attr : currentECLInclude)
|
{
|
if (!newECLInclude.contains(attr))
|
{
|
configECLIncludeChanged = true;
|
break;
|
}
|
}
|
}
|
|
if (configECLIncludeChanged)
|
{
|
// set new config
|
this.setEclInclude(this.serverID, newECLInclude);
|
if (broker != null)
|
{
|
disableService();
|
enableService();
|
}
|
}
|
}
|
|
/**
|
* This method should trigger an export of the replicated data.
|
* to the provided outputStream.
|
* When finished the outputStream should be flushed and closed.
|
*
|
* @param output The OutputStream where the export should
|
* be produced.
|
* @throws DirectoryException When needed.
|
*/
|
abstract protected void exportBackend(OutputStream output)
|
throws DirectoryException;
|
|
/**
|
* This method should trigger an import of the replicated data.
|
*
|
* @param input The InputStream from which
|
* the import should be reading entries.
|
*
|
* @throws DirectoryException When needed.
|
*/
|
abstract protected void importBackend(InputStream input)
|
throws DirectoryException;
|
|
/**
|
* This method should return the total number of objects in the
|
* replicated domain.
|
* This count will be used for reporting.
|
*
|
* @throws DirectoryException when needed.
|
*
|
* @return The number of objects in the replication domain.
|
*/
|
public abstract long countEntries() throws DirectoryException;
|
|
/**
|
* This method should handle the processing of {@link UpdateMsg} receive
|
* from remote replication entities.
|
* <p>
|
* This method will be called by a single thread and should therefore
|
* should not be blocking.
|
*
|
* @param updateMsg The {@link UpdateMsg} that was received.
|
*
|
* @return A boolean indicating if the processing is completed at return
|
* time.
|
* If <code> true </code> is returned, no further
|
* processing is necessary.
|
*
|
* If <code> false </code> is returned, the subclass should
|
* call the method
|
* {@link #processUpdateDone(UpdateMsg, String)}
|
* and update the ServerState
|
* When this processing is complete.
|
*
|
*/
|
public abstract boolean processUpdate(UpdateMsg updateMsg);
|
|
/**
|
* This method must be called after each call to
|
* {@link #processUpdate(UpdateMsg)} when the processing of the update is
|
* completed.
|
* <p>
|
* It is useful for implementation needing to process the update in an
|
* asynchronous way or using several threads, but must be called even
|
* by implementation doing it in a synchronous, single-threaded way.
|
*
|
* @param msg The UpdateMsg whose processing was completed.
|
* @param replayErrorMsg if not null, this means an error occurred during the
|
* replay of this update, and this is the matching human readable message
|
* describing the problem.
|
*/
|
public void processUpdateDone(UpdateMsg msg, String replayErrorMsg)
|
{
|
broker.updateWindowAfterReplay();
|
|
// Send an ack if it was requested and the group id is the same of the RS
|
// one. Only Safe Read mode makes sense in DS for returning an ack.
|
byte rsGroupId = broker.getRsGroupId();
|
if (msg.isAssured())
|
{
|
// Assured feature is supported starting from replication protocol V2
|
if (broker.getProtocolVersion() >=
|
ProtocolVersion.REPLICATION_PROTOCOL_V2)
|
{
|
AssuredMode msgAssuredMode = msg.getAssuredMode();
|
if (msgAssuredMode == AssuredMode.SAFE_READ_MODE)
|
{
|
if (rsGroupId == groupId)
|
{
|
// Send the ack
|
AckMsg ackMsg = new AckMsg(msg.getChangeNumber());
|
if (replayErrorMsg != null)
|
{
|
// Mark the error in the ack
|
// -> replay error occured
|
ackMsg.setHasReplayError(true);
|
// -> replay error occured in our server
|
List<Integer> idList = new ArrayList<Integer>();
|
idList.add(serverID);
|
ackMsg.setFailedServers(idList);
|
}
|
broker.publish(ackMsg);
|
if (replayErrorMsg != null)
|
{
|
assuredSrReceivedUpdatesNotAcked.incrementAndGet();
|
} else
|
{
|
assuredSrReceivedUpdatesAcked.incrementAndGet();
|
}
|
}
|
} else if (assuredMode != AssuredMode.SAFE_DATA_MODE)
|
{
|
Message errorMsg = ERR_DS_UNKNOWN_ASSURED_MODE.get(
|
Integer.toString(serverID), msgAssuredMode.toString(), serviceID,
|
msg.toString());
|
logError(errorMsg);
|
} else
|
{
|
// In safe data mode assured update that comes up to a DS requires no
|
// ack from a destinator DS. Safe data mode is based on RS acks only
|
}
|
}
|
}
|
|
incProcessedUpdates();
|
}
|
|
/**
|
* Prepare a message if it is to be sent in assured mode.
|
* If the assured mode is enabled, this method should be called before
|
* publish(UpdateMsg msg) method. This will configure the update accordingly
|
* before it is sent and will prepare the mechanism that will block until the
|
* matching ack is received. To wait for the ack after publish call, use
|
* the waitForAckIfAssuredEnabled() method.
|
* The expected typical usage in a service inheriting from this class is
|
* the following sequence:
|
* UpdateMsg msg = xxx;
|
* prepareWaitForAckIfAssuredEnabled(msg);
|
* publish(msg);
|
* waitForAckIfAssuredEnabled(msg);
|
*
|
* Note: prepareWaitForAckIfAssuredEnabled and waitForAckIfAssuredEnabled have
|
* no effect if assured replication is disabled.
|
* Note: this mechanism should not be used if using publish(byte[] msg)
|
* version as usage of these methods is already hidden inside.
|
*
|
* @param msg The update message to be sent soon.
|
*/
|
protected void prepareWaitForAckIfAssuredEnabled(UpdateMsg msg)
|
{
|
byte rsGroupId = broker.getRsGroupId();
|
/*
|
* If assured configured, set message accordingly to request an ack in the
|
* right assured mode.
|
* No ack requested for a RS with a different group id. Assured
|
* replication suported for the same locality, i.e: a topology working in
|
* the same
|
* geographical location). If we are connected to a RS which is not in our
|
* locality, no need to ask for an ack.
|
*/
|
if (assured && (rsGroupId == groupId))
|
{
|
msg.setAssured(true);
|
msg.setAssuredMode(assuredMode);
|
if (assuredMode == AssuredMode.SAFE_DATA_MODE)
|
msg.setSafeDataLevel(assuredSdLevel);
|
|
// Add the assured message to the list of update that are
|
// waiting for acks
|
synchronized (waitingAckMsgs)
|
{
|
waitingAckMsgs.put(msg.getChangeNumber(), msg);
|
}
|
}
|
}
|
|
/**
|
* Wait for the processing of an assured message after it has been sent, if
|
* assured replication is configured, otherwise, do nothing.
|
* The prepareWaitForAckIfAssuredEnabled method should have been called
|
* before, see its comment for the full picture.
|
*
|
* @param msg The UpdateMsg for which we are waiting for an ack.
|
* @throws TimeoutException When the configured timeout occurs waiting for the
|
* ack.
|
*/
|
protected void waitForAckIfAssuredEnabled(UpdateMsg msg)
|
throws TimeoutException
|
{
|
byte rsGroupId = broker.getRsGroupId();
|
|
// If assured mode configured, wait for acknowledgement for the just sent
|
// message
|
if (assured && (rsGroupId == groupId))
|
{
|
// Increment assured replication monitoring counters
|
switch (assuredMode)
|
{
|
case SAFE_READ_MODE:
|
assuredSrSentUpdates.incrementAndGet();
|
break;
|
case SAFE_DATA_MODE:
|
assuredSdSentUpdates.incrementAndGet();
|
break;
|
default:
|
// Should not happen
|
}
|
} else
|
{
|
// Not assured or bad group id, return immediately
|
return;
|
}
|
|
// Wait for the ack to be received, timing out if necessary
|
long startTime = System.currentTimeMillis();
|
synchronized (msg)
|
{
|
ChangeNumber cn = msg.getChangeNumber();
|
while (waitingAckMsgs.containsKey(cn))
|
{
|
try
|
{
|
// WARNING: this timeout may be difficult to optimize: too low, it
|
// may use too much CPU, too high, it may penalize performance...
|
msg.wait(10);
|
} catch (InterruptedException e)
|
{
|
if (debugEnabled())
|
{
|
TRACER.debugInfo("waitForAck method interrupted for replication " +
|
"serviceID: " + serviceID);
|
}
|
break;
|
}
|
// Timeout ?
|
if ( (System.currentTimeMillis() - startTime) >= assuredTimeout )
|
{
|
// Timeout occured, be sure that ack is not being received and if so,
|
// remove the update from the wait list, log the timeout error and
|
// also update assured monitoring counters
|
UpdateMsg update;
|
synchronized (waitingAckMsgs)
|
{
|
update = waitingAckMsgs.remove(cn);
|
}
|
|
if (update != null)
|
{
|
// No luck, this is a real timeout
|
// Increment assured replication monitoring counters
|
switch (msg.getAssuredMode())
|
{
|
case SAFE_READ_MODE:
|
assuredSrNotAcknowledgedUpdates.incrementAndGet();
|
assuredSrTimeoutUpdates.incrementAndGet();
|
// Increment number of errors for our RS
|
updateAssuredErrorsByServer(
|
assuredSrServerNotAcknowledgedUpdates,
|
broker.getRsServerId());
|
break;
|
case SAFE_DATA_MODE:
|
assuredSdTimeoutUpdates.incrementAndGet();
|
// Increment number of errors for our RS
|
updateAssuredErrorsByServer(assuredSdServerTimeoutUpdates,
|
broker.getRsServerId());
|
break;
|
default:
|
// Should not happen
|
}
|
|
throw new TimeoutException("No ack received for message cn: " + cn +
|
" and replication servceID: " + serviceID + " after " +
|
assuredTimeout + " ms.");
|
} else
|
{
|
// Ack received just before timeout limit: we can exit
|
break;
|
}
|
}
|
}
|
}
|
}
|
|
/**
|
* Publish an {@link UpdateMsg} to the Replication Service.
|
* <p>
|
* The Replication Service will handle the delivery of this {@link UpdateMsg}
|
* to all the participants of this Replication Domain.
|
* These members will be receive this {@link UpdateMsg} through a call
|
* of the {@link #processUpdate(UpdateMsg)} message.
|
*
|
* @param msg The UpdateMsg that should be pushed.
|
*/
|
public void publish(UpdateMsg msg)
|
{
|
// Publish the update
|
broker.publish(msg);
|
state.update(msg.getChangeNumber());
|
numSentUpdates.incrementAndGet();
|
}
|
|
/**
|
* Publish informations to the Replication Service (not assured mode).
|
*
|
* @param msg The byte array containing the informations that should
|
* be sent to the remote entities.
|
*/
|
public void publish(byte[] msg)
|
{
|
UpdateMsg update;
|
synchronized (this)
|
{
|
update = new UpdateMsg(generator.newChangeNumber(), msg);
|
|
// If assured replication is configured, this will prepare blocking
|
// mechanism. If assured replication is disabled, this returns
|
// immediately
|
prepareWaitForAckIfAssuredEnabled(update);
|
|
publish(update);
|
}
|
|
try
|
{
|
// If assured replication is enabled, this will wait for the matching
|
// ack or time out. If assured replication is disabled, this returns
|
// immediately
|
waitForAckIfAssuredEnabled(update);
|
} catch (TimeoutException ex)
|
{
|
// This exception may only be raised if assured replication is
|
// enabled
|
Message errorMsg = NOTE_DS_ACK_TIMEOUT.get(serviceID, Long.toString(
|
assuredTimeout), msg.toString());
|
logError(errorMsg);
|
}
|
}
|
|
/**
|
* This method should return the generationID to use for this
|
* ReplicationDomain.
|
* This method can be called at any time after the ReplicationDomain
|
* has been started.
|
*
|
* @return The GenerationID.
|
*/
|
public abstract long getGenerationID();
|
|
/**
|
* Subclasses should use this method to add additional monitoring
|
* information in the ReplicationDomain.
|
*
|
* @return Additional monitoring attributes that will be added in the
|
* ReplicationDomain monitoring entry.
|
*/
|
public Collection<Attribute> getAdditionalMonitoring()
|
{
|
return new ArrayList<Attribute>();
|
}
|
|
/**
|
* Returns a boolean indicating if a total update import is currently
|
* in Progress.
|
*
|
* @return A boolean indicating if a total update import is currently
|
* in Progress.
|
*/
|
public boolean importInProgress()
|
{
|
if (ieContext == null)
|
return false;
|
else
|
return ieContext.importInProgress;
|
}
|
|
/**
|
* Returns a boolean indicating if a total update export is currently
|
* in Progress.
|
*
|
* @return A boolean indicating if a total update export is currently
|
* in Progress.
|
*/
|
public boolean exportInProgress()
|
{
|
if (ieContext == null)
|
return false;
|
else
|
return !ieContext.importInProgress;
|
}
|
|
/**
|
* Returns the number of entries still to be processed when a total update
|
* is in progress.
|
*
|
* @return The number of entries still to be processed when a total update
|
* is in progress.
|
*/
|
long getLeftEntryCount()
|
{
|
if (ieContext != null)
|
return ieContext.entryLeftCount;
|
else
|
return 0;
|
}
|
|
/**
|
* Returns the total number of entries to be processed when a total update
|
* is in progress.
|
*
|
* @return The total number of entries to be processed when a total update
|
* is in progress.
|
*/
|
long getTotalEntryCount()
|
{
|
if (ieContext != null)
|
return ieContext.entryCount;
|
else
|
return 0;
|
}
|
|
/**
|
* Set the attributes configured on a server to be included in the ECL.
|
* @param serverId server where these attributes are configured.
|
* @param attributes the configured attributes.
|
*/
|
public void setEclInclude(int serverId, Set<String> attributes)
|
{
|
synchronized(eclIncludeByServer)
|
{
|
eclIncludeByServer.put(serverId, attributes);
|
|
// and rebuild the global list to be ready for usage
|
crossServersECLIncludes.clear();
|
for (Set<String> attributesByServer : eclIncludeByServer.values())
|
for (String attribute : attributesByServer)
|
crossServersECLIncludes.add(attribute);
|
}
|
}
|
|
/**
|
* Get the attributes to include in each change for the ECL.
|
* It's a set : an attribute appears once even if configured on more than one
|
* server.
|
* @return The attributes to include in each change for the ECL.
|
*/
|
public Set<String> getEclInclude()
|
{
|
return crossServersECLIncludes;
|
}
|
|
/**
|
* Get the attributes to include in each change for the ECL
|
* for a given serverId.
|
* @param serverId The serverId for which we want the include attributes.
|
* @return The attributes.
|
*/
|
public Set<String> getEclInclude(int serverId)
|
{
|
synchronized(eclIncludeByServer)
|
{
|
return eclIncludeByServer.get(serverId);
|
}
|
}
|
|
/**
|
* Returns the ChangeNUmber of the last Change that was fully processed
|
* by this ReplicationDomain.
|
*
|
* @return The ChangeNUmber of the last Change that was fully processed
|
* by this ReplicationDomain.
|
*/
|
public ChangeNumber getLastLocalChange()
|
{
|
return state.getMaxChangeNumber(serverID);
|
}
|
}
|