/**
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.camel.processor;
import java.util.ArrayList;
import java.util.Collection;
import java.util.Iterator;
import java.util.LinkedList;
import java.util.List;
import java.util.Queue;
import java.util.concurrent.TimeUnit;
import java.util.concurrent.locks.Condition;
import java.util.concurrent.locks.Lock;
import java.util.concurrent.locks.ReentrantLock;
import org.apache.camel.Exchange;
import org.apache.camel.Navigate;
import org.apache.camel.Processor;
import org.apache.camel.impl.GroupedExchange;
import org.apache.camel.impl.LoggingExceptionHandler;
import org.apache.camel.impl.ServiceSupport;
import org.apache.camel.spi.ExceptionHandler;
import org.apache.camel.util.ObjectHelper;
import org.apache.camel.util.ServiceHelper;
import org.apache.camel.util.concurrent.ExecutorServiceHelper;
import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory;
/**
* A base class for any kind of {@link Processor} which implements some kind of batch processing.
*
* @version $Revision: 782594 $
*/
public class BatchProcessor extends ServiceSupport implements Processor, Navigate<Processor> {
public static final long DEFAULT_BATCH_TIMEOUT = 1000L;
public static final int DEFAULT_BATCH_SIZE = 100;
private static final Log LOG = LogFactory.getLog(BatchProcessor.class);
private long batchTimeout = DEFAULT_BATCH_TIMEOUT;
private int batchSize = DEFAULT_BATCH_SIZE;
private int outBatchSize;
private boolean groupExchanges;
private boolean batchConsumer;
private final Processor processor;
private final Collection<Exchange> collection;
private ExceptionHandler exceptionHandler;
private final BatchSender sender;
public BatchProcessor(Processor processor, Collection<Exchange> collection) {
ObjectHelper.notNull(processor, "processor");
ObjectHelper.notNull(collection, "collection");
this.processor = processor;
this.collection = collection;
this.sender = new BatchSender();
}
@Override
public String toString() {
return "BatchProcessor[to: " + processor + "]";
}
// Properties
// -------------------------------------------------------------------------
public ExceptionHandler getExceptionHandler() {
if (exceptionHandler == null) {
exceptionHandler = new LoggingExceptionHandler(getClass());
}
return exceptionHandler;
}
public void setExceptionHandler(ExceptionHandler exceptionHandler) {
this.exceptionHandler = exceptionHandler;
}
public int getBatchSize() {
return batchSize;
}
/**
* Sets the <b>in</b> batch size. This is the number of incoming exchanges that this batch processor will
* process before its completed. The default value is {@link #DEFAULT_BATCH_SIZE}.
*
* @param batchSize the size
*/
public void setBatchSize(int batchSize) {
// setting batch size to 0 or negative is like disabling it, so we set it as the max value
// as the code logic is dependt on a batch size having 1..n value
if (batchSize <= 0) {
LOG.debug("Disabling batch size, will only be triggered by timeout");
this.batchSize = Integer.MAX_VALUE;
} else {
this.batchSize = batchSize;
}
}
public int getOutBatchSize() {
return outBatchSize;
}
/**
* Sets the <b>out</b> batch size. If the batch processor holds more exchanges than this out size then the
* completion is triggered. Can for instance be used to ensure that this batch is completed when a certain
* number of exchanges has been collected. By default this feature is <b>not</b> enabled.
*
* @param outBatchSize the size
*/
public void setOutBatchSize(int outBatchSize) {
this.outBatchSize = outBatchSize;
}
public long getBatchTimeout() {
return batchTimeout;
}
public void setBatchTimeout(long batchTimeout) {
this.batchTimeout = batchTimeout;
}
public boolean isGroupExchanges() {
return groupExchanges;
}
public void setGroupExchanges(boolean groupExchanges) {
this.groupExchanges = groupExchanges;
}
public boolean isBatchConsumer() {
return batchConsumer;
}
public void setBatchConsumer(boolean batchConsumer) {
this.batchConsumer = batchConsumer;
}
public Processor getProcessor() {
return processor;
}
public List<Processor> next() {
if (!hasNext()) {
return null;
}
List<Processor> answer = new ArrayList<Processor>(1);
answer.add(processor);
return answer;
}
public boolean hasNext() {
return processor != null;
}
/**
* A strategy method to decide if the "in" batch is completed. That is, whether the resulting exchanges in
* the in queue should be drained to the "out" collection.
*/
private boolean isInBatchCompleted(int num) {
return num >= batchSize;
}
/**
* A strategy method to decide if the "out" batch is completed. That is, whether the resulting exchange in
* the out collection should be sent.
*/
private boolean isOutBatchCompleted() {
if (outBatchSize == 0) {
// out batch is disabled, so go ahead and send.
return true;
}
return collection.size() > 0 && collection.size() >= outBatchSize;
}
/**
* Strategy Method to process an exchange in the batch. This method allows derived classes to perform
* custom processing before or after an individual exchange is processed
*/
protected void processExchange(Exchange exchange) throws Exception {
processor.process(exchange);
}
protected void doStart() throws Exception {
ServiceHelper.startServices(processor);
sender.start();
}
protected void doStop() throws Exception {
sender.cancel();
ServiceHelper.stopServices(sender);
ServiceHelper.stopServices(processor);
collection.clear();
}
/**
* Enqueues an exchange for later batch processing.
*/
public void process(Exchange exchange) throws Exception {
// if batch consumer is enabled then we need to adjust the batch size
// with the size from the batch consumer
if (isBatchConsumer()) {
int size = exchange.getProperty(Exchange.BATCH_SIZE, Integer.class);
if (batchSize != size) {
batchSize = size;
if (LOG.isTraceEnabled()) {
LOG.trace("Using batch consumer completion, so setting batch size to: " + batchSize);
}
}
}
sender.enqueueExchange(exchange);
}
/**
* Sender thread for queued-up exchanges.
*/
private class BatchSender extends Thread {
private Queue<Exchange> queue;
private Lock queueLock = new ReentrantLock();
private boolean exchangeEnqueued;
private Condition exchangeEnqueuedCondition = queueLock.newCondition();
public BatchSender() {
super(ExecutorServiceHelper.getThreadName("Batch Sender"));
this.queue = new LinkedList<Exchange>();
}
@Override
public void run() {
// Wait until one of either:
// * an exchange being queued;
// * the batch timeout expiring; or
// * the thread being cancelled.
//
// If an exchange is queued then we need to determine whether the
// batch is complete. If it is complete then we send out the batched
// exchanges. Otherwise we move back into our wait state.
//
// If the batch times out then we send out the batched exchanges
// collected so far.
//
// If we receive an interrupt then all blocking operations are
// interrupted and our thread terminates.
//
// The goal of the following algorithm in terms of synchronisation
// is to provide fine grained locking i.e. retaining the lock only
// when required. Special consideration is given to releasing the
// lock when calling an overloaded method i.e. sendExchanges.
// Unlocking is important as the process of sending out the exchanges
// would otherwise block new exchanges from being queued.
queueLock.lock();
try {
do {
try {
if (!exchangeEnqueued) {
exchangeEnqueuedCondition.await(batchTimeout, TimeUnit.MILLISECONDS);
}
if (!exchangeEnqueued) {
drainQueueTo(collection, batchSize);
} else {
exchangeEnqueued = false;
while (isInBatchCompleted(queue.size())) {
drainQueueTo(collection, batchSize);
}
if (!isOutBatchCompleted()) {
continue;
}
}
queueLock.unlock();
try {
try {
sendExchanges();
} catch (Exception e) {
getExceptionHandler().handleException(e);
}
} finally {
queueLock.lock();
}
} catch (InterruptedException e) {
break;
}
} while (isRunAllowed());
} finally {
queueLock.unlock();
}
}
/**
* This method should be called with queueLock held
*/
private void drainQueueTo(Collection<Exchange> collection, int batchSize) {
for (int i = 0; i < batchSize; ++i) {
Exchange e = queue.poll();
if (e != null) {
collection.add(e);
} else {
break;
}
}
}
public void cancel() {
interrupt();
}
public void enqueueExchange(Exchange exchange) {
queueLock.lock();
try {
queue.add(exchange);
exchangeEnqueued = true;
exchangeEnqueuedCondition.signal();
} finally {
queueLock.unlock();
}
}
private void sendExchanges() throws Exception {
GroupedExchange grouped = null;
Iterator<Exchange> iter = collection.iterator();
while (iter.hasNext()) {
Exchange exchange = iter.next();
iter.remove();
if (!groupExchanges) {
// non grouped so process the exchange one at a time
processExchange(exchange);
} else {
// grouped so add all exchanges into one group
if (grouped == null) {
grouped = new GroupedExchange(exchange.getContext());
}
grouped.addExchange(exchange);
}
}
// and after adding process the single grouped exchange
if (grouped != null) {
processExchange(grouped);
}
}
}
}