Merge 816b55ffb0 into 9720b1b249
This commit is contained in:
commit
f211be88d3
29 changed files with 2094 additions and 355 deletions
2
pom.xml
2
pom.xml
|
|
@ -6,7 +6,7 @@
|
||||||
<artifactId>amazon-kinesis-client</artifactId>
|
<artifactId>amazon-kinesis-client</artifactId>
|
||||||
<packaging>jar</packaging>
|
<packaging>jar</packaging>
|
||||||
<name>Amazon Kinesis Client Library for Java</name>
|
<name>Amazon Kinesis Client Library for Java</name>
|
||||||
<version>1.8.5</version>
|
<version>1.8.6-SNAPSHOT</version>
|
||||||
<description>The Amazon Kinesis Client Library for Java enables Java developers to easily consume and process data
|
<description>The Amazon Kinesis Client Library for Java enables Java developers to easily consume and process data
|
||||||
from Amazon Kinesis.
|
from Amazon Kinesis.
|
||||||
</description>
|
</description>
|
||||||
|
|
|
||||||
|
|
@ -0,0 +1,90 @@
|
||||||
|
/*
|
||||||
|
* Copyright 2017 Amazon.com, Inc. or its affiliates. All Rights Reserved.
|
||||||
|
*
|
||||||
|
* Licensed under the Amazon Software License (the "License").
|
||||||
|
* You may not use this file except in compliance with the License.
|
||||||
|
* A copy of the License is located at
|
||||||
|
*
|
||||||
|
* http://aws.amazon.com/asl/
|
||||||
|
*
|
||||||
|
* or in the "license" file accompanying this file. This file is distributed
|
||||||
|
* on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
|
||||||
|
* express or implied. See the License for the specific language governing
|
||||||
|
* permissions and limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package com.amazonaws.services.kinesis.clientlibrary.lib.worker;
|
||||||
|
|
||||||
|
import java.time.Duration;
|
||||||
|
import java.time.Instant;
|
||||||
|
|
||||||
|
import com.amazonaws.services.kinesis.clientlibrary.types.ProcessRecordsInput;
|
||||||
|
import com.amazonaws.services.kinesis.model.GetRecordsResult;
|
||||||
|
|
||||||
|
import lombok.extern.apachecommons.CommonsLog;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* This is the BlockingGetRecordsCache class. This class blocks any calls to the getRecords on the
|
||||||
|
* GetRecordsRetrievalStrategy class.
|
||||||
|
*/
|
||||||
|
@CommonsLog
|
||||||
|
public class BlockingGetRecordsCache implements GetRecordsCache {
|
||||||
|
private final int maxRecordsPerCall;
|
||||||
|
private final GetRecordsRetrievalStrategy getRecordsRetrievalStrategy;
|
||||||
|
private final long idleMillisBetweenCalls;
|
||||||
|
private Instant lastSuccessfulCall;
|
||||||
|
|
||||||
|
public BlockingGetRecordsCache(final int maxRecordsPerCall,
|
||||||
|
final GetRecordsRetrievalStrategy getRecordsRetrievalStrategy,
|
||||||
|
final long idleMillisBetweenCalls) {
|
||||||
|
this.maxRecordsPerCall = maxRecordsPerCall;
|
||||||
|
this.getRecordsRetrievalStrategy = getRecordsRetrievalStrategy;
|
||||||
|
this.idleMillisBetweenCalls = idleMillisBetweenCalls;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void start() {
|
||||||
|
//
|
||||||
|
// Nothing to do here
|
||||||
|
//
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public ProcessRecordsInput getNextResult() {
|
||||||
|
sleepBeforeNextCall();
|
||||||
|
GetRecordsResult getRecordsResult = getRecordsRetrievalStrategy.getRecords(maxRecordsPerCall);
|
||||||
|
lastSuccessfulCall = Instant.now();
|
||||||
|
ProcessRecordsInput processRecordsInput = new ProcessRecordsInput()
|
||||||
|
.withRecords(getRecordsResult.getRecords())
|
||||||
|
.withMillisBehindLatest(getRecordsResult.getMillisBehindLatest());
|
||||||
|
return processRecordsInput;
|
||||||
|
}
|
||||||
|
|
||||||
|
private void sleepBeforeNextCall() {
|
||||||
|
if (!Thread.interrupted()) {
|
||||||
|
if (lastSuccessfulCall == null) {
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
long timeSinceLastCall = Duration.between(lastSuccessfulCall, Instant.now()).abs().toMillis();
|
||||||
|
if (timeSinceLastCall < idleMillisBetweenCalls) {
|
||||||
|
try {
|
||||||
|
Thread.sleep(idleMillisBetweenCalls - timeSinceLastCall);
|
||||||
|
} catch (InterruptedException e) {
|
||||||
|
log.info("Thread was interrupted, indicating that shutdown was called.");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
log.info("Thread has been interrupted, indicating that it is in the shutdown phase.");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public GetRecordsRetrievalStrategy getGetRecordsRetrievalStrategy() {
|
||||||
|
return getRecordsRetrievalStrategy;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void shutdown() {
|
||||||
|
getRecordsRetrievalStrategy.shutdown();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
@ -14,8 +14,6 @@
|
||||||
*/
|
*/
|
||||||
package com.amazonaws.services.kinesis.clientlibrary.lib.worker;
|
package com.amazonaws.services.kinesis.clientlibrary.lib.worker;
|
||||||
|
|
||||||
import java.util.Optional;
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Top level container for all the possible states a {@link ShardConsumer} can be in. The logic for creation of tasks,
|
* Top level container for all the possible states a {@link ShardConsumer} can be in. The logic for creation of tasks,
|
||||||
* and state transitions is contained within the {@link ConsumerState} objects.
|
* and state transitions is contained within the {@link ConsumerState} objects.
|
||||||
|
|
@ -253,9 +251,14 @@ class ConsumerStates {
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public ITask createTask(ShardConsumer consumer) {
|
public ITask createTask(ShardConsumer consumer) {
|
||||||
return new InitializeTask(consumer.getShardInfo(), consumer.getRecordProcessor(), consumer.getCheckpoint(),
|
return new InitializeTask(consumer.getShardInfo(),
|
||||||
consumer.getRecordProcessorCheckpointer(), consumer.getDataFetcher(),
|
consumer.getRecordProcessor(),
|
||||||
consumer.getTaskBackoffTimeMillis(), consumer.getStreamConfig());
|
consumer.getCheckpoint(),
|
||||||
|
consumer.getRecordProcessorCheckpointer(),
|
||||||
|
consumer.getDataFetcher(),
|
||||||
|
consumer.getTaskBackoffTimeMillis(),
|
||||||
|
consumer.getStreamConfig(),
|
||||||
|
consumer.getGetRecordsCache());
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
|
@ -309,10 +312,14 @@ class ConsumerStates {
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public ITask createTask(ShardConsumer consumer) {
|
public ITask createTask(ShardConsumer consumer) {
|
||||||
return new ProcessTask(consumer.getShardInfo(), consumer.getStreamConfig(), consumer.getRecordProcessor(),
|
return new ProcessTask(consumer.getShardInfo(),
|
||||||
consumer.getRecordProcessorCheckpointer(), consumer.getDataFetcher(),
|
consumer.getStreamConfig(),
|
||||||
consumer.getTaskBackoffTimeMillis(), consumer.isSkipShardSyncAtWorkerInitializationIfLeasesExist(),
|
consumer.getRecordProcessor(),
|
||||||
consumer.getGetRecordsRetrievalStrategy());
|
consumer.getRecordProcessorCheckpointer(),
|
||||||
|
consumer.getDataFetcher(),
|
||||||
|
consumer.getTaskBackoffTimeMillis(),
|
||||||
|
consumer.isSkipShardSyncAtWorkerInitializationIfLeasesExist(),
|
||||||
|
consumer.getGetRecordsCache());
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
|
@ -371,8 +378,10 @@ class ConsumerStates {
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public ITask createTask(ShardConsumer consumer) {
|
public ITask createTask(ShardConsumer consumer) {
|
||||||
return new ShutdownNotificationTask(consumer.getRecordProcessor(), consumer.getRecordProcessorCheckpointer(),
|
return new ShutdownNotificationTask(consumer.getRecordProcessor(),
|
||||||
consumer.getShutdownNotification(), consumer.getShardInfo());
|
consumer.getRecordProcessorCheckpointer(),
|
||||||
|
consumer.getShutdownNotification(),
|
||||||
|
consumer.getShardInfo());
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
|
@ -511,13 +520,16 @@ class ConsumerStates {
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public ITask createTask(ShardConsumer consumer) {
|
public ITask createTask(ShardConsumer consumer) {
|
||||||
return new ShutdownTask(consumer.getShardInfo(), consumer.getRecordProcessor(),
|
return new ShutdownTask(consumer.getShardInfo(),
|
||||||
consumer.getRecordProcessorCheckpointer(), consumer.getShutdownReason(),
|
consumer.getRecordProcessor(),
|
||||||
|
consumer.getRecordProcessorCheckpointer(),
|
||||||
|
consumer.getShutdownReason(),
|
||||||
consumer.getStreamConfig().getStreamProxy(),
|
consumer.getStreamConfig().getStreamProxy(),
|
||||||
consumer.getStreamConfig().getInitialPositionInStream(),
|
consumer.getStreamConfig().getInitialPositionInStream(),
|
||||||
consumer.isCleanupLeasesOfCompletedShards(), consumer.getLeaseManager(),
|
consumer.isCleanupLeasesOfCompletedShards(),
|
||||||
|
consumer.getLeaseManager(),
|
||||||
consumer.getTaskBackoffTimeMillis(),
|
consumer.getTaskBackoffTimeMillis(),
|
||||||
consumer.getGetRecordsRetrievalStrategy());
|
consumer.getGetRecordsCache());
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
|
|
||||||
|
|
@ -0,0 +1,8 @@
|
||||||
|
package com.amazonaws.services.kinesis.clientlibrary.lib.worker;
|
||||||
|
|
||||||
|
/**
|
||||||
|
*
|
||||||
|
*/
|
||||||
|
public enum DataFetchingStrategy {
|
||||||
|
DEFAULT, PREFETCH_CACHED;
|
||||||
|
}
|
||||||
|
|
@ -0,0 +1,43 @@
|
||||||
|
/*
|
||||||
|
* Copyright 2017 Amazon.com, Inc. or its affiliates. All Rights Reserved.
|
||||||
|
*
|
||||||
|
* Licensed under the Amazon Software License (the "License").
|
||||||
|
* You may not use this file except in compliance with the License.
|
||||||
|
* A copy of the License is located at
|
||||||
|
*
|
||||||
|
* http://aws.amazon.com/asl/
|
||||||
|
*
|
||||||
|
* or in the "license" file accompanying this file. This file is distributed
|
||||||
|
* on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
|
||||||
|
* express or implied. See the License for the specific language governing
|
||||||
|
* permissions and limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package com.amazonaws.services.kinesis.clientlibrary.lib.worker;
|
||||||
|
|
||||||
|
import com.amazonaws.services.kinesis.clientlibrary.types.ProcessRecordsInput;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* This class is used as a cache for Prefetching data from Kinesis.
|
||||||
|
*/
|
||||||
|
public interface GetRecordsCache {
|
||||||
|
/**
|
||||||
|
* This method calls the start behavior on the cache, if available.
|
||||||
|
*/
|
||||||
|
void start();
|
||||||
|
|
||||||
|
/**
|
||||||
|
* This method returns the next set of records from the Cache if present, or blocks the request till it gets the
|
||||||
|
* next set of records back from Kinesis.
|
||||||
|
*
|
||||||
|
* @return The next set of records.
|
||||||
|
*/
|
||||||
|
ProcessRecordsInput getNextResult();
|
||||||
|
|
||||||
|
GetRecordsRetrievalStrategy getGetRecordsRetrievalStrategy();
|
||||||
|
|
||||||
|
/**
|
||||||
|
* This method calls the shutdown behavior on the cache, if available.
|
||||||
|
*/
|
||||||
|
void shutdown();
|
||||||
|
}
|
||||||
|
|
@ -0,0 +1,12 @@
|
||||||
|
package com.amazonaws.services.kinesis.clientlibrary.lib.worker;
|
||||||
|
|
||||||
|
import com.amazonaws.services.kinesis.model.GetRecordsResult;
|
||||||
|
|
||||||
|
import java.util.concurrent.Callable;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* This class uses the GetRecordsRetrievalStrategy class to retrieve the next set of records and update the cache.
|
||||||
|
*/
|
||||||
|
public interface GetRecordsRetriever {
|
||||||
|
GetRecordsResult getNextRecords(int maxRecords);
|
||||||
|
}
|
||||||
|
|
@ -1,16 +1,16 @@
|
||||||
/*
|
/*
|
||||||
* Copyright 2012-2016 Amazon.com, Inc. or its affiliates. All Rights Reserved.
|
* Copyright 2017 Amazon.com, Inc. or its affiliates. All Rights Reserved.
|
||||||
*
|
*
|
||||||
* Licensed under the Amazon Software License (the "License").
|
* Licensed under the Amazon Software License (the "License").
|
||||||
* You may not use this file except in compliance with the License.
|
* You may not use this file except in compliance with the License.
|
||||||
* A copy of the License is located at
|
* A copy of the License is located at
|
||||||
*
|
*
|
||||||
* http://aws.amazon.com/asl/
|
* http://aws.amazon.com/asl/
|
||||||
*
|
*
|
||||||
* or in the "license" file accompanying this file. This file is distributed
|
* or in the "license" file accompanying this file. This file is distributed
|
||||||
* on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
|
* on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
|
||||||
* express or implied. See the License for the specific language governing
|
* express or implied. See the License for the specific language governing
|
||||||
* permissions and limitations under the License.
|
* permissions and limitations under the License.
|
||||||
*/
|
*/
|
||||||
package com.amazonaws.services.kinesis.clientlibrary.lib.worker;
|
package com.amazonaws.services.kinesis.clientlibrary.lib.worker;
|
||||||
|
|
||||||
|
|
@ -43,17 +43,19 @@ class InitializeTask implements ITask {
|
||||||
// Back off for this interval if we encounter a problem (exception)
|
// Back off for this interval if we encounter a problem (exception)
|
||||||
private final long backoffTimeMillis;
|
private final long backoffTimeMillis;
|
||||||
private final StreamConfig streamConfig;
|
private final StreamConfig streamConfig;
|
||||||
|
private final GetRecordsCache getRecordsCache;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Constructor.
|
* Constructor.
|
||||||
*/
|
*/
|
||||||
InitializeTask(ShardInfo shardInfo,
|
InitializeTask(ShardInfo shardInfo,
|
||||||
IRecordProcessor recordProcessor,
|
IRecordProcessor recordProcessor,
|
||||||
ICheckpoint checkpoint,
|
ICheckpoint checkpoint,
|
||||||
RecordProcessorCheckpointer recordProcessorCheckpointer,
|
RecordProcessorCheckpointer recordProcessorCheckpointer,
|
||||||
KinesisDataFetcher dataFetcher,
|
KinesisDataFetcher dataFetcher,
|
||||||
long backoffTimeMillis,
|
long backoffTimeMillis,
|
||||||
StreamConfig streamConfig) {
|
StreamConfig streamConfig,
|
||||||
|
GetRecordsCache getRecordsCache) {
|
||||||
this.shardInfo = shardInfo;
|
this.shardInfo = shardInfo;
|
||||||
this.recordProcessor = recordProcessor;
|
this.recordProcessor = recordProcessor;
|
||||||
this.checkpoint = checkpoint;
|
this.checkpoint = checkpoint;
|
||||||
|
|
@ -61,6 +63,7 @@ class InitializeTask implements ITask {
|
||||||
this.dataFetcher = dataFetcher;
|
this.dataFetcher = dataFetcher;
|
||||||
this.backoffTimeMillis = backoffTimeMillis;
|
this.backoffTimeMillis = backoffTimeMillis;
|
||||||
this.streamConfig = streamConfig;
|
this.streamConfig = streamConfig;
|
||||||
|
this.getRecordsCache = getRecordsCache;
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
|
@ -80,6 +83,7 @@ class InitializeTask implements ITask {
|
||||||
ExtendedSequenceNumber initialCheckpoint = initialCheckpointObject.getCheckpoint();
|
ExtendedSequenceNumber initialCheckpoint = initialCheckpointObject.getCheckpoint();
|
||||||
|
|
||||||
dataFetcher.initialize(initialCheckpoint.getSequenceNumber(), streamConfig.getInitialPositionInStream());
|
dataFetcher.initialize(initialCheckpoint.getSequenceNumber(), streamConfig.getInitialPositionInStream());
|
||||||
|
getRecordsCache.start();
|
||||||
recordProcessorCheckpointer.setLargestPermittedCheckpointValue(initialCheckpoint);
|
recordProcessorCheckpointer.setLargestPermittedCheckpointValue(initialCheckpoint);
|
||||||
recordProcessorCheckpointer.setInitialCheckpointValue(initialCheckpoint);
|
recordProcessorCheckpointer.setInitialCheckpointValue(initialCheckpoint);
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -1,16 +1,16 @@
|
||||||
/*
|
/*
|
||||||
* Copyright 2012-2016 Amazon.com, Inc. or its affiliates. All Rights Reserved.
|
* Copyright 2017 Amazon.com, Inc. or its affiliates. All Rights Reserved.
|
||||||
*
|
*
|
||||||
* Licensed under the Amazon Software License (the "License").
|
* Licensed under the Amazon Software License (the "License").
|
||||||
* You may not use this file except in compliance with the License.
|
* You may not use this file except in compliance with the License.
|
||||||
* A copy of the License is located at
|
* A copy of the License is located at
|
||||||
*
|
*
|
||||||
* http://aws.amazon.com/asl/
|
* http://aws.amazon.com/asl/
|
||||||
*
|
*
|
||||||
* or in the "license" file accompanying this file. This file is distributed
|
* or in the "license" file accompanying this file. This file is distributed
|
||||||
* on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
|
* on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
|
||||||
* express or implied. See the License for the specific language governing
|
* express or implied. See the License for the specific language governing
|
||||||
* permissions and limitations under the License.
|
* permissions and limitations under the License.
|
||||||
*/
|
*/
|
||||||
package com.amazonaws.services.kinesis.clientlibrary.lib.worker;
|
package com.amazonaws.services.kinesis.clientlibrary.lib.worker;
|
||||||
|
|
||||||
|
|
@ -232,6 +232,9 @@ public class KinesisClientLibConfiguration {
|
||||||
@Getter
|
@Getter
|
||||||
private int maxLeaseRenewalThreads = DEFAULT_MAX_LEASE_RENEWAL_THREADS;
|
private int maxLeaseRenewalThreads = DEFAULT_MAX_LEASE_RENEWAL_THREADS;
|
||||||
|
|
||||||
|
@Getter
|
||||||
|
private RecordsFetcherFactory recordsFetcherFactory;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Constructor.
|
* Constructor.
|
||||||
*
|
*
|
||||||
|
|
@ -267,14 +270,30 @@ public class KinesisClientLibConfiguration {
|
||||||
AWSCredentialsProvider dynamoDBCredentialsProvider,
|
AWSCredentialsProvider dynamoDBCredentialsProvider,
|
||||||
AWSCredentialsProvider cloudWatchCredentialsProvider,
|
AWSCredentialsProvider cloudWatchCredentialsProvider,
|
||||||
String workerId) {
|
String workerId) {
|
||||||
this(applicationName, streamName, null, null, DEFAULT_INITIAL_POSITION_IN_STREAM, kinesisCredentialsProvider,
|
this(applicationName,
|
||||||
dynamoDBCredentialsProvider, cloudWatchCredentialsProvider, DEFAULT_FAILOVER_TIME_MILLIS, workerId,
|
streamName,
|
||||||
DEFAULT_MAX_RECORDS, DEFAULT_IDLETIME_BETWEEN_READS_MILLIS,
|
null,
|
||||||
DEFAULT_DONT_CALL_PROCESS_RECORDS_FOR_EMPTY_RECORD_LIST, DEFAULT_PARENT_SHARD_POLL_INTERVAL_MILLIS,
|
null,
|
||||||
DEFAULT_SHARD_SYNC_INTERVAL_MILLIS, DEFAULT_CLEANUP_LEASES_UPON_SHARDS_COMPLETION,
|
DEFAULT_INITIAL_POSITION_IN_STREAM,
|
||||||
new ClientConfiguration(), new ClientConfiguration(), new ClientConfiguration(),
|
kinesisCredentialsProvider,
|
||||||
DEFAULT_TASK_BACKOFF_TIME_MILLIS, DEFAULT_METRICS_BUFFER_TIME_MILLIS, DEFAULT_METRICS_MAX_QUEUE_SIZE,
|
dynamoDBCredentialsProvider,
|
||||||
DEFAULT_VALIDATE_SEQUENCE_NUMBER_BEFORE_CHECKPOINTING, null,
|
cloudWatchCredentialsProvider,
|
||||||
|
DEFAULT_FAILOVER_TIME_MILLIS,
|
||||||
|
workerId,
|
||||||
|
DEFAULT_MAX_RECORDS,
|
||||||
|
DEFAULT_IDLETIME_BETWEEN_READS_MILLIS,
|
||||||
|
DEFAULT_DONT_CALL_PROCESS_RECORDS_FOR_EMPTY_RECORD_LIST,
|
||||||
|
DEFAULT_PARENT_SHARD_POLL_INTERVAL_MILLIS,
|
||||||
|
DEFAULT_SHARD_SYNC_INTERVAL_MILLIS,
|
||||||
|
DEFAULT_CLEANUP_LEASES_UPON_SHARDS_COMPLETION,
|
||||||
|
new ClientConfiguration(),
|
||||||
|
new ClientConfiguration(),
|
||||||
|
new ClientConfiguration(),
|
||||||
|
DEFAULT_TASK_BACKOFF_TIME_MILLIS,
|
||||||
|
DEFAULT_METRICS_BUFFER_TIME_MILLIS,
|
||||||
|
DEFAULT_METRICS_MAX_QUEUE_SIZE,
|
||||||
|
DEFAULT_VALIDATE_SEQUENCE_NUMBER_BEFORE_CHECKPOINTING,
|
||||||
|
null,
|
||||||
DEFAULT_SHUTDOWN_GRACE_MILLIS);
|
DEFAULT_SHUTDOWN_GRACE_MILLIS);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -315,29 +334,29 @@ public class KinesisClientLibConfiguration {
|
||||||
// CHECKSTYLE:IGNORE HiddenFieldCheck FOR NEXT 26 LINES
|
// CHECKSTYLE:IGNORE HiddenFieldCheck FOR NEXT 26 LINES
|
||||||
// CHECKSTYLE:IGNORE ParameterNumber FOR NEXT 26 LINES
|
// CHECKSTYLE:IGNORE ParameterNumber FOR NEXT 26 LINES
|
||||||
public KinesisClientLibConfiguration(String applicationName,
|
public KinesisClientLibConfiguration(String applicationName,
|
||||||
String streamName,
|
String streamName,
|
||||||
String kinesisEndpoint,
|
String kinesisEndpoint,
|
||||||
InitialPositionInStream initialPositionInStream,
|
InitialPositionInStream initialPositionInStream,
|
||||||
AWSCredentialsProvider kinesisCredentialsProvider,
|
AWSCredentialsProvider kinesisCredentialsProvider,
|
||||||
AWSCredentialsProvider dynamoDBCredentialsProvider,
|
AWSCredentialsProvider dynamoDBCredentialsProvider,
|
||||||
AWSCredentialsProvider cloudWatchCredentialsProvider,
|
AWSCredentialsProvider cloudWatchCredentialsProvider,
|
||||||
long failoverTimeMillis,
|
long failoverTimeMillis,
|
||||||
String workerId,
|
String workerId,
|
||||||
int maxRecords,
|
int maxRecords,
|
||||||
long idleTimeBetweenReadsInMillis,
|
long idleTimeBetweenReadsInMillis,
|
||||||
boolean callProcessRecordsEvenForEmptyRecordList,
|
boolean callProcessRecordsEvenForEmptyRecordList,
|
||||||
long parentShardPollIntervalMillis,
|
long parentShardPollIntervalMillis,
|
||||||
long shardSyncIntervalMillis,
|
long shardSyncIntervalMillis,
|
||||||
boolean cleanupTerminatedShardsBeforeExpiry,
|
boolean cleanupTerminatedShardsBeforeExpiry,
|
||||||
ClientConfiguration kinesisClientConfig,
|
ClientConfiguration kinesisClientConfig,
|
||||||
ClientConfiguration dynamoDBClientConfig,
|
ClientConfiguration dynamoDBClientConfig,
|
||||||
ClientConfiguration cloudWatchClientConfig,
|
ClientConfiguration cloudWatchClientConfig,
|
||||||
long taskBackoffTimeMillis,
|
long taskBackoffTimeMillis,
|
||||||
long metricsBufferTimeMillis,
|
long metricsBufferTimeMillis,
|
||||||
int metricsMaxQueueSize,
|
int metricsMaxQueueSize,
|
||||||
boolean validateSequenceNumberBeforeCheckpointing,
|
boolean validateSequenceNumberBeforeCheckpointing,
|
||||||
String regionName,
|
String regionName,
|
||||||
long shutdownGraceMillis) {
|
long shutdownGraceMillis) {
|
||||||
this(applicationName, streamName, kinesisEndpoint, null, initialPositionInStream, kinesisCredentialsProvider,
|
this(applicationName, streamName, kinesisEndpoint, null, initialPositionInStream, kinesisCredentialsProvider,
|
||||||
dynamoDBCredentialsProvider, cloudWatchCredentialsProvider, failoverTimeMillis, workerId,
|
dynamoDBCredentialsProvider, cloudWatchCredentialsProvider, failoverTimeMillis, workerId,
|
||||||
maxRecords, idleTimeBetweenReadsInMillis,
|
maxRecords, idleTimeBetweenReadsInMillis,
|
||||||
|
|
@ -345,7 +364,117 @@ public class KinesisClientLibConfiguration {
|
||||||
shardSyncIntervalMillis, cleanupTerminatedShardsBeforeExpiry,
|
shardSyncIntervalMillis, cleanupTerminatedShardsBeforeExpiry,
|
||||||
kinesisClientConfig, dynamoDBClientConfig, cloudWatchClientConfig,
|
kinesisClientConfig, dynamoDBClientConfig, cloudWatchClientConfig,
|
||||||
taskBackoffTimeMillis, metricsBufferTimeMillis, metricsMaxQueueSize,
|
taskBackoffTimeMillis, metricsBufferTimeMillis, metricsMaxQueueSize,
|
||||||
validateSequenceNumberBeforeCheckpointing, regionName, shutdownGraceMillis);
|
validateSequenceNumberBeforeCheckpointing, regionName, shutdownGraceMillis);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @param applicationName Name of the Kinesis application
|
||||||
|
* By default the application name is included in the user agent string used to make AWS requests. This
|
||||||
|
* can assist with troubleshooting (e.g. distinguish requests made by separate applications).
|
||||||
|
* @param streamName Name of the Kinesis stream
|
||||||
|
* @param kinesisEndpoint Kinesis endpoint
|
||||||
|
* @param dynamoDBEndpoint DynamoDB endpoint
|
||||||
|
* @param initialPositionInStream One of LATEST or TRIM_HORIZON. The KinesisClientLibrary will start fetching
|
||||||
|
* records from that location in the stream when an application starts up for the first time and there
|
||||||
|
* are no checkpoints. If there are checkpoints, then we start from the checkpoint position.
|
||||||
|
* @param kinesisCredentialsProvider Provides credentials used to access Kinesis
|
||||||
|
* @param dynamoDBCredentialsProvider Provides credentials used to access DynamoDB
|
||||||
|
* @param cloudWatchCredentialsProvider Provides credentials used to access CloudWatch
|
||||||
|
* @param failoverTimeMillis Lease duration (leases not renewed within this period will be claimed by others)
|
||||||
|
* @param workerId Used to distinguish different workers/processes of a Kinesis application
|
||||||
|
* @param maxRecords Max records to read per Kinesis getRecords() call
|
||||||
|
* @param idleTimeBetweenReadsInMillis Idle time between calls to fetch data from Kinesis
|
||||||
|
* @param callProcessRecordsEvenForEmptyRecordList Call the IRecordProcessor::processRecords() API even if
|
||||||
|
* GetRecords returned an empty record list.
|
||||||
|
* @param parentShardPollIntervalMillis Wait for this long between polls to check if parent shards are done
|
||||||
|
* @param shardSyncIntervalMillis Time between tasks to sync leases and Kinesis shards
|
||||||
|
* @param cleanupTerminatedShardsBeforeExpiry Clean up shards we've finished processing (don't wait for expiration
|
||||||
|
* in Kinesis)
|
||||||
|
* @param kinesisClientConfig Client Configuration used by Kinesis client
|
||||||
|
* @param dynamoDBClientConfig Client Configuration used by DynamoDB client
|
||||||
|
* @param cloudWatchClientConfig Client Configuration used by CloudWatch client
|
||||||
|
* @param taskBackoffTimeMillis Backoff period when tasks encounter an exception
|
||||||
|
* @param metricsBufferTimeMillis Metrics are buffered for at most this long before publishing to CloudWatch
|
||||||
|
* @param metricsMaxQueueSize Max number of metrics to buffer before publishing to CloudWatch
|
||||||
|
* @param validateSequenceNumberBeforeCheckpointing whether KCL should validate client provided sequence numbers
|
||||||
|
* with a call to Amazon Kinesis before checkpointing for calls to
|
||||||
|
* {@link RecordProcessorCheckpointer#checkpoint(String)}
|
||||||
|
* @param regionName The region name for the service
|
||||||
|
*/
|
||||||
|
// CHECKSTYLE:IGNORE HiddenFieldCheck FOR NEXT 26 LINES
|
||||||
|
// CHECKSTYLE:IGNORE ParameterNumber FOR NEXT 26 LINES
|
||||||
|
public KinesisClientLibConfiguration(String applicationName,
|
||||||
|
String streamName,
|
||||||
|
String kinesisEndpoint,
|
||||||
|
String dynamoDBEndpoint,
|
||||||
|
InitialPositionInStream initialPositionInStream,
|
||||||
|
AWSCredentialsProvider kinesisCredentialsProvider,
|
||||||
|
AWSCredentialsProvider dynamoDBCredentialsProvider,
|
||||||
|
AWSCredentialsProvider cloudWatchCredentialsProvider,
|
||||||
|
long failoverTimeMillis,
|
||||||
|
String workerId,
|
||||||
|
int maxRecords,
|
||||||
|
long idleTimeBetweenReadsInMillis,
|
||||||
|
boolean callProcessRecordsEvenForEmptyRecordList,
|
||||||
|
long parentShardPollIntervalMillis,
|
||||||
|
long shardSyncIntervalMillis,
|
||||||
|
boolean cleanupTerminatedShardsBeforeExpiry,
|
||||||
|
ClientConfiguration kinesisClientConfig,
|
||||||
|
ClientConfiguration dynamoDBClientConfig,
|
||||||
|
ClientConfiguration cloudWatchClientConfig,
|
||||||
|
long taskBackoffTimeMillis,
|
||||||
|
long metricsBufferTimeMillis,
|
||||||
|
int metricsMaxQueueSize,
|
||||||
|
boolean validateSequenceNumberBeforeCheckpointing,
|
||||||
|
String regionName,
|
||||||
|
long shutdownGraceMillis) {
|
||||||
|
// Check following values are greater than zero
|
||||||
|
checkIsValuePositive("FailoverTimeMillis", failoverTimeMillis);
|
||||||
|
checkIsValuePositive("IdleTimeBetweenReadsInMillis", idleTimeBetweenReadsInMillis);
|
||||||
|
checkIsValuePositive("ParentShardPollIntervalMillis", parentShardPollIntervalMillis);
|
||||||
|
checkIsValuePositive("ShardSyncIntervalMillis", shardSyncIntervalMillis);
|
||||||
|
checkIsValuePositive("MaxRecords", (long) maxRecords);
|
||||||
|
checkIsValuePositive("TaskBackoffTimeMillis", taskBackoffTimeMillis);
|
||||||
|
checkIsValuePositive("MetricsBufferTimeMills", metricsBufferTimeMillis);
|
||||||
|
checkIsValuePositive("MetricsMaxQueueSize", (long) metricsMaxQueueSize);
|
||||||
|
checkIsValuePositive("ShutdownGraceMillis", shutdownGraceMillis);
|
||||||
|
checkIsRegionNameValid(regionName);
|
||||||
|
this.applicationName = applicationName;
|
||||||
|
this.tableName = applicationName;
|
||||||
|
this.streamName = streamName;
|
||||||
|
this.kinesisEndpoint = kinesisEndpoint;
|
||||||
|
this.dynamoDBEndpoint = dynamoDBEndpoint;
|
||||||
|
this.initialPositionInStream = initialPositionInStream;
|
||||||
|
this.kinesisCredentialsProvider = kinesisCredentialsProvider;
|
||||||
|
this.dynamoDBCredentialsProvider = dynamoDBCredentialsProvider;
|
||||||
|
this.cloudWatchCredentialsProvider = cloudWatchCredentialsProvider;
|
||||||
|
this.failoverTimeMillis = failoverTimeMillis;
|
||||||
|
this.maxRecords = maxRecords;
|
||||||
|
this.idleTimeBetweenReadsInMillis = idleTimeBetweenReadsInMillis;
|
||||||
|
this.callProcessRecordsEvenForEmptyRecordList = callProcessRecordsEvenForEmptyRecordList;
|
||||||
|
this.parentShardPollIntervalMillis = parentShardPollIntervalMillis;
|
||||||
|
this.shardSyncIntervalMillis = shardSyncIntervalMillis;
|
||||||
|
this.cleanupLeasesUponShardCompletion = cleanupTerminatedShardsBeforeExpiry;
|
||||||
|
this.workerIdentifier = workerId;
|
||||||
|
this.kinesisClientConfig = checkAndAppendKinesisClientLibUserAgent(kinesisClientConfig);
|
||||||
|
this.dynamoDBClientConfig = checkAndAppendKinesisClientLibUserAgent(dynamoDBClientConfig);
|
||||||
|
this.cloudWatchClientConfig = checkAndAppendKinesisClientLibUserAgent(cloudWatchClientConfig);
|
||||||
|
this.taskBackoffTimeMillis = taskBackoffTimeMillis;
|
||||||
|
this.metricsBufferTimeMillis = metricsBufferTimeMillis;
|
||||||
|
this.metricsMaxQueueSize = metricsMaxQueueSize;
|
||||||
|
this.metricsLevel = DEFAULT_METRICS_LEVEL;
|
||||||
|
this.metricsEnabledDimensions = DEFAULT_METRICS_ENABLED_DIMENSIONS;
|
||||||
|
this.validateSequenceNumberBeforeCheckpointing = validateSequenceNumberBeforeCheckpointing;
|
||||||
|
this.regionName = regionName;
|
||||||
|
this.maxLeasesForWorker = DEFAULT_MAX_LEASES_FOR_WORKER;
|
||||||
|
this.maxLeasesToStealAtOneTime = DEFAULT_MAX_LEASES_TO_STEAL_AT_ONE_TIME;
|
||||||
|
this.initialLeaseTableReadCapacity = DEFAULT_INITIAL_LEASE_TABLE_READ_CAPACITY;
|
||||||
|
this.initialLeaseTableWriteCapacity = DEFAULT_INITIAL_LEASE_TABLE_WRITE_CAPACITY;
|
||||||
|
this.initialPositionInStreamExtended =
|
||||||
|
InitialPositionInStreamExtended.newInitialPosition(initialPositionInStream);
|
||||||
|
this.skipShardSyncAtWorkerInitializationIfLeasesExist = DEFAULT_SKIP_SHARD_SYNC_AT_STARTUP_IF_LEASES_EXIST;
|
||||||
|
this.shardPrioritization = DEFAULT_SHARD_PRIORITIZATION;
|
||||||
|
this.recordsFetcherFactory = new SimpleRecordsFetcherFactory(this.maxRecords);
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
|
@ -408,7 +537,7 @@ public class KinesisClientLibConfiguration {
|
||||||
int metricsMaxQueueSize,
|
int metricsMaxQueueSize,
|
||||||
boolean validateSequenceNumberBeforeCheckpointing,
|
boolean validateSequenceNumberBeforeCheckpointing,
|
||||||
String regionName,
|
String regionName,
|
||||||
long shutdownGraceMillis) {
|
RecordsFetcherFactory recordsFetcherFactory) {
|
||||||
// Check following values are greater than zero
|
// Check following values are greater than zero
|
||||||
checkIsValuePositive("FailoverTimeMillis", failoverTimeMillis);
|
checkIsValuePositive("FailoverTimeMillis", failoverTimeMillis);
|
||||||
checkIsValuePositive("IdleTimeBetweenReadsInMillis", idleTimeBetweenReadsInMillis);
|
checkIsValuePositive("IdleTimeBetweenReadsInMillis", idleTimeBetweenReadsInMillis);
|
||||||
|
|
@ -418,7 +547,6 @@ public class KinesisClientLibConfiguration {
|
||||||
checkIsValuePositive("TaskBackoffTimeMillis", taskBackoffTimeMillis);
|
checkIsValuePositive("TaskBackoffTimeMillis", taskBackoffTimeMillis);
|
||||||
checkIsValuePositive("MetricsBufferTimeMills", metricsBufferTimeMillis);
|
checkIsValuePositive("MetricsBufferTimeMills", metricsBufferTimeMillis);
|
||||||
checkIsValuePositive("MetricsMaxQueueSize", (long) metricsMaxQueueSize);
|
checkIsValuePositive("MetricsMaxQueueSize", (long) metricsMaxQueueSize);
|
||||||
checkIsValuePositive("ShutdownGraceMillis", shutdownGraceMillis);
|
|
||||||
checkIsRegionNameValid(regionName);
|
checkIsRegionNameValid(regionName);
|
||||||
this.applicationName = applicationName;
|
this.applicationName = applicationName;
|
||||||
this.tableName = applicationName;
|
this.tableName = applicationName;
|
||||||
|
|
@ -455,6 +583,7 @@ public class KinesisClientLibConfiguration {
|
||||||
InitialPositionInStreamExtended.newInitialPosition(initialPositionInStream);
|
InitialPositionInStreamExtended.newInitialPosition(initialPositionInStream);
|
||||||
this.skipShardSyncAtWorkerInitializationIfLeasesExist = DEFAULT_SKIP_SHARD_SYNC_AT_STARTUP_IF_LEASES_EXIST;
|
this.skipShardSyncAtWorkerInitializationIfLeasesExist = DEFAULT_SKIP_SHARD_SYNC_AT_STARTUP_IF_LEASES_EXIST;
|
||||||
this.shardPrioritization = DEFAULT_SHARD_PRIORITIZATION;
|
this.shardPrioritization = DEFAULT_SHARD_PRIORITIZATION;
|
||||||
|
this.recordsFetcherFactory = recordsFetcherFactory;
|
||||||
this.shutdownGraceMillis = shutdownGraceMillis;
|
this.shutdownGraceMillis = shutdownGraceMillis;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -1158,6 +1287,48 @@ public class KinesisClientLibConfiguration {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
*
|
||||||
|
* @param maxPendingProcessRecordsInput The max number of ProcessRecordsInput that can be stored in the cache before
|
||||||
|
* blocking
|
||||||
|
* @return this configuration object
|
||||||
|
*/
|
||||||
|
public KinesisClientLibConfiguration withMaxPendingProcessRecordsInput(final int maxPendingProcessRecordsInput) {
|
||||||
|
checkIsValuePositive("maxPendingProcessRecordsInput", maxPendingProcessRecordsInput);
|
||||||
|
this.recordsFetcherFactory.setMaxPendingProcessRecordsInput(maxPendingProcessRecordsInput);
|
||||||
|
return this;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @param maxCacheByteSize Max byte size for the cache at any given point of time. After this threshold is crossed
|
||||||
|
* the KinesisDataFetcher will be blocked until the cache has more space available.
|
||||||
|
* @return KinesisClientLibConfiguration
|
||||||
|
*/
|
||||||
|
public KinesisClientLibConfiguration withMaxCacheByteSize(final int maxCacheByteSize) {
|
||||||
|
checkIsValuePositive("maxCacheByteSize", maxCacheByteSize);
|
||||||
|
this.recordsFetcherFactory.setMaxByteSize(maxCacheByteSize);
|
||||||
|
return this;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @param dataFetchingStrategy The strategy for fetching data from kinesis.
|
||||||
|
* @return KinesisClientLibConfiguration
|
||||||
|
*/
|
||||||
|
public KinesisClientLibConfiguration withDataFetchingStrategy(String dataFetchingStrategy) {
|
||||||
|
this.recordsFetcherFactory.setDataFetchingStrategy(DataFetchingStrategy.valueOf(dataFetchingStrategy.toUpperCase()));
|
||||||
|
return this;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @param maxRecordsCount The maximum number of records in the cache, accross all ProcessRecordInput objects
|
||||||
|
* @return KinesisClientLibConfiguration
|
||||||
|
*/
|
||||||
|
public KinesisClientLibConfiguration withMaxRecordsCount(final int maxRecordsCount) {
|
||||||
|
checkIsValuePositive("maxRecordsCount", maxRecordsCount);
|
||||||
|
this.recordsFetcherFactory.setMaxRecordsCount(maxRecordsCount);
|
||||||
|
return this;
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* @param timeoutInSeconds The timeout in seconds to wait for the MultiLangProtocol to wait for
|
* @param timeoutInSeconds The timeout in seconds to wait for the MultiLangProtocol to wait for
|
||||||
*/
|
*/
|
||||||
|
|
@ -1174,4 +1345,14 @@ public class KinesisClientLibConfiguration {
|
||||||
this.shutdownGraceMillis = shutdownGraceMillis;
|
this.shutdownGraceMillis = shutdownGraceMillis;
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @param idleMillisBetweenCalls Idle time between 2 getcalls from the data fetcher.
|
||||||
|
* @return KinesisClientLibConfiguration
|
||||||
|
*/
|
||||||
|
public KinesisClientLibConfiguration withIdleMillisBetweenCalls(long idleMillisBetweenCalls) {
|
||||||
|
checkIsValuePositive("IdleMillisBetweenCalls", idleMillisBetweenCalls);
|
||||||
|
this.recordsFetcherFactory.setIdleMillisBetweenCalls(idleMillisBetweenCalls);
|
||||||
|
return this;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -1,33 +1,34 @@
|
||||||
/*
|
/*
|
||||||
* Copyright 2012-2016 Amazon.com, Inc. or its affiliates. All Rights Reserved.
|
* Copyright 2017 Amazon.com, Inc. or its affiliates. All Rights Reserved.
|
||||||
*
|
*
|
||||||
* Licensed under the Amazon Software License (the "License").
|
* Licensed under the Amazon Software License (the "License").
|
||||||
* You may not use this file except in compliance with the License.
|
* You may not use this file except in compliance with the License.
|
||||||
* A copy of the License is located at
|
* A copy of the License is located at
|
||||||
*
|
*
|
||||||
* http://aws.amazon.com/asl/
|
* http://aws.amazon.com/asl/
|
||||||
*
|
*
|
||||||
* or in the "license" file accompanying this file. This file is distributed
|
* or in the "license" file accompanying this file. This file is distributed
|
||||||
* on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
|
* on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
|
||||||
* express or implied. See the License for the specific language governing
|
* express or implied. See the License for the specific language governing
|
||||||
* permissions and limitations under the License.
|
* permissions and limitations under the License.
|
||||||
*/
|
*/
|
||||||
package com.amazonaws.services.kinesis.clientlibrary.lib.worker;
|
package com.amazonaws.services.kinesis.clientlibrary.lib.worker;
|
||||||
|
|
||||||
import lombok.Data;
|
import java.util.Collections;
|
||||||
|
import java.util.Date;
|
||||||
|
|
||||||
import org.apache.commons.logging.Log;
|
import org.apache.commons.logging.Log;
|
||||||
import org.apache.commons.logging.LogFactory;
|
import org.apache.commons.logging.LogFactory;
|
||||||
|
|
||||||
import com.amazonaws.services.kinesis.model.GetRecordsResult;
|
|
||||||
import com.amazonaws.services.kinesis.model.ResourceNotFoundException;
|
|
||||||
import com.amazonaws.services.kinesis.model.ShardIteratorType;
|
|
||||||
import com.amazonaws.services.kinesis.clientlibrary.lib.checkpoint.SentinelCheckpoint;
|
import com.amazonaws.services.kinesis.clientlibrary.lib.checkpoint.SentinelCheckpoint;
|
||||||
import com.amazonaws.services.kinesis.clientlibrary.proxies.IKinesisProxy;
|
import com.amazonaws.services.kinesis.clientlibrary.proxies.IKinesisProxy;
|
||||||
import com.amazonaws.services.kinesis.clientlibrary.proxies.MetricsCollectingKinesisProxyDecorator;
|
import com.amazonaws.services.kinesis.clientlibrary.proxies.MetricsCollectingKinesisProxyDecorator;
|
||||||
import com.amazonaws.services.kinesis.clientlibrary.types.ExtendedSequenceNumber;
|
import com.amazonaws.services.kinesis.clientlibrary.types.ExtendedSequenceNumber;
|
||||||
|
import com.amazonaws.services.kinesis.model.GetRecordsResult;
|
||||||
|
import com.amazonaws.services.kinesis.model.ResourceNotFoundException;
|
||||||
|
import com.amazonaws.services.kinesis.model.ShardIteratorType;
|
||||||
|
|
||||||
import java.util.Date;
|
import lombok.Data;
|
||||||
import java.util.function.Consumer;
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Used to get data from Amazon Kinesis. Tracks iterator state internally.
|
* Used to get data from Amazon Kinesis. Tracks iterator state internally.
|
||||||
|
|
@ -49,8 +50,7 @@ class KinesisDataFetcher {
|
||||||
*/
|
*/
|
||||||
public KinesisDataFetcher(IKinesisProxy kinesisProxy, ShardInfo shardInfo) {
|
public KinesisDataFetcher(IKinesisProxy kinesisProxy, ShardInfo shardInfo) {
|
||||||
this.shardId = shardInfo.getShardId();
|
this.shardId = shardInfo.getShardId();
|
||||||
this.kinesisProxy =
|
this.kinesisProxy = new MetricsCollectingKinesisProxyDecorator("KinesisDataFetcher", kinesisProxy, this.shardId);
|
||||||
new MetricsCollectingKinesisProxyDecorator("KinesisDataFetcher", kinesisProxy, this.shardId);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
|
@ -64,25 +64,23 @@ class KinesisDataFetcher {
|
||||||
throw new IllegalArgumentException("KinesisDataFetcher.getRecords called before initialization.");
|
throw new IllegalArgumentException("KinesisDataFetcher.getRecords called before initialization.");
|
||||||
}
|
}
|
||||||
|
|
||||||
DataFetcherResult response;
|
|
||||||
if (nextIterator != null) {
|
if (nextIterator != null) {
|
||||||
try {
|
try {
|
||||||
response = new AdvancingResult(kinesisProxy.get(nextIterator, maxRecords));
|
return new AdvancingResult(kinesisProxy.get(nextIterator, maxRecords));
|
||||||
} catch (ResourceNotFoundException e) {
|
} catch (ResourceNotFoundException e) {
|
||||||
LOG.info("Caught ResourceNotFoundException when fetching records for shard " + shardId);
|
LOG.info("Caught ResourceNotFoundException when fetching records for shard " + shardId);
|
||||||
response = TERMINAL_RESULT;
|
return TERMINAL_RESULT;
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
response = TERMINAL_RESULT;
|
return TERMINAL_RESULT;
|
||||||
}
|
}
|
||||||
|
|
||||||
return response;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
final DataFetcherResult TERMINAL_RESULT = new DataFetcherResult() {
|
final DataFetcherResult TERMINAL_RESULT = new DataFetcherResult() {
|
||||||
@Override
|
@Override
|
||||||
public GetRecordsResult getResult() {
|
public GetRecordsResult getResult() {
|
||||||
return null;
|
return new GetRecordsResult().withMillisBehindLatest(null).withRecords(Collections.emptyList())
|
||||||
|
.withNextShardIterator(null);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
|
@ -98,7 +96,7 @@ class KinesisDataFetcher {
|
||||||
};
|
};
|
||||||
|
|
||||||
@Data
|
@Data
|
||||||
private class AdvancingResult implements DataFetcherResult {
|
class AdvancingResult implements DataFetcherResult {
|
||||||
|
|
||||||
final GetRecordsResult result;
|
final GetRecordsResult result;
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -0,0 +1,223 @@
|
||||||
|
/*
|
||||||
|
* Copyright 2017 Amazon.com, Inc. or its affiliates. All Rights Reserved.
|
||||||
|
*
|
||||||
|
* Licensed under the Amazon Software License (the "License").
|
||||||
|
* You may not use this file except in compliance with the License.
|
||||||
|
* A copy of the License is located at
|
||||||
|
*
|
||||||
|
* http://aws.amazon.com/asl/
|
||||||
|
*
|
||||||
|
* or in the "license" file accompanying this file. This file is distributed
|
||||||
|
* on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
|
||||||
|
* express or implied. See the License for the specific language governing
|
||||||
|
* permissions and limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package com.amazonaws.services.kinesis.clientlibrary.lib.worker;
|
||||||
|
|
||||||
|
import java.time.Duration;
|
||||||
|
import java.time.Instant;
|
||||||
|
import java.util.concurrent.ExecutorService;
|
||||||
|
import java.util.concurrent.LinkedBlockingQueue;
|
||||||
|
|
||||||
|
import com.amazonaws.SdkClientException;
|
||||||
|
import com.amazonaws.services.kinesis.clientlibrary.types.ProcessRecordsInput;
|
||||||
|
import com.amazonaws.services.kinesis.metrics.impl.MetricsHelper;
|
||||||
|
import com.amazonaws.services.kinesis.metrics.impl.ThreadSafeMetricsDelegatingFactory;
|
||||||
|
import com.amazonaws.services.kinesis.metrics.interfaces.IMetricsFactory;
|
||||||
|
import com.amazonaws.services.kinesis.model.GetRecordsResult;
|
||||||
|
|
||||||
|
import lombok.NonNull;
|
||||||
|
import lombok.extern.apachecommons.CommonsLog;
|
||||||
|
import org.apache.commons.lang.Validate;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* This is the prefetch caching class, this class spins up a thread if prefetching is enabled. That thread fetches the
|
||||||
|
* next set of records and stores it in the cache. The size of the cache is limited by setting
|
||||||
|
* maxPendingProcessRecordsInput i.e. the maximum number of GetRecordsResult that the cache can store, maxByteSize
|
||||||
|
* i.e. the byte size of the records stored in the cache and maxRecordsCount i.e. the max number of records that should
|
||||||
|
* be present in the cache across multiple GetRecordsResult object. If no data is available in the cache, the call from
|
||||||
|
* the record processor is blocked till records are retrieved from Kinesis.
|
||||||
|
*/
|
||||||
|
@CommonsLog
|
||||||
|
public class PrefetchGetRecordsCache implements GetRecordsCache {
|
||||||
|
LinkedBlockingQueue<ProcessRecordsInput> getRecordsResultQueue;
|
||||||
|
private int maxPendingProcessRecordsInput;
|
||||||
|
private int maxByteSize;
|
||||||
|
private int maxRecordsCount;
|
||||||
|
private final int maxRecordsPerCall;
|
||||||
|
private final GetRecordsRetrievalStrategy getRecordsRetrievalStrategy;
|
||||||
|
private final ExecutorService executorService;
|
||||||
|
private final IMetricsFactory metricsFactory;
|
||||||
|
private final long idleMillisBetweenCalls;
|
||||||
|
private Instant lastSuccessfulCall;
|
||||||
|
private final DefaultGetRecordsCacheDaemon defaultGetRecordsCacheDaemon;
|
||||||
|
private PrefetchCounters prefetchCounters;
|
||||||
|
private boolean started = false;
|
||||||
|
private final String operation;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Constructor for the PrefetchGetRecordsCache. This cache prefetches records from Kinesis and stores them in a
|
||||||
|
* LinkedBlockingQueue.
|
||||||
|
*
|
||||||
|
* @see com.amazonaws.services.kinesis.clientlibrary.lib.worker.PrefetchGetRecordsCache
|
||||||
|
*
|
||||||
|
* @param maxPendingProcessRecordsInput Max number of ProcessRecordsInput that can be held in the cache before
|
||||||
|
* blocking
|
||||||
|
* @param maxByteSize Max byte size of the queue before blocking next get records call
|
||||||
|
* @param maxRecordsCount Max number of records in the queue across all ProcessRecordInput objects
|
||||||
|
* @param maxRecordsPerCall Max records to be returned per call
|
||||||
|
* @param getRecordsRetrievalStrategy Retrieval strategy for the get records call
|
||||||
|
* @param executorService Executor service for the cache
|
||||||
|
* @param idleMillisBetweenCalls maximum time to wait before dispatching the next get records call
|
||||||
|
*/
|
||||||
|
public PrefetchGetRecordsCache(final int maxPendingProcessRecordsInput, final int maxByteSize, final int maxRecordsCount,
|
||||||
|
final int maxRecordsPerCall,
|
||||||
|
@NonNull final GetRecordsRetrievalStrategy getRecordsRetrievalStrategy,
|
||||||
|
@NonNull final ExecutorService executorService,
|
||||||
|
long idleMillisBetweenCalls,
|
||||||
|
@NonNull final IMetricsFactory metricsFactory,
|
||||||
|
@NonNull String operation) {
|
||||||
|
this.getRecordsRetrievalStrategy = getRecordsRetrievalStrategy;
|
||||||
|
this.maxRecordsPerCall = maxRecordsPerCall;
|
||||||
|
this.maxPendingProcessRecordsInput = maxPendingProcessRecordsInput;
|
||||||
|
this.maxByteSize = maxByteSize;
|
||||||
|
this.maxRecordsCount = maxRecordsCount;
|
||||||
|
this.getRecordsResultQueue = new LinkedBlockingQueue<>(this.maxPendingProcessRecordsInput);
|
||||||
|
this.prefetchCounters = new PrefetchCounters();
|
||||||
|
this.executorService = executorService;
|
||||||
|
this.metricsFactory = new ThreadSafeMetricsDelegatingFactory(metricsFactory);
|
||||||
|
this.idleMillisBetweenCalls = idleMillisBetweenCalls;
|
||||||
|
this.defaultGetRecordsCacheDaemon = new DefaultGetRecordsCacheDaemon();
|
||||||
|
Validate.notEmpty(operation, "Operation cannot be empty");
|
||||||
|
this.operation = operation;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void start() {
|
||||||
|
if (executorService.isShutdown()) {
|
||||||
|
throw new IllegalStateException("ExecutorService has been shutdown.");
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!started) {
|
||||||
|
log.info("Starting prefetching thread.");
|
||||||
|
executorService.execute(defaultGetRecordsCacheDaemon);
|
||||||
|
}
|
||||||
|
started = true;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public ProcessRecordsInput getNextResult() {
|
||||||
|
if (executorService.isShutdown()) {
|
||||||
|
throw new IllegalStateException("Shutdown has been called on the cache, can't accept new requests.");
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!started) {
|
||||||
|
throw new IllegalStateException("Cache has not been initialized, make sure to call start.");
|
||||||
|
}
|
||||||
|
ProcessRecordsInput result = null;
|
||||||
|
try {
|
||||||
|
result = getRecordsResultQueue.take().withCacheExitTime(Instant.now());
|
||||||
|
prefetchCounters.removed(result);
|
||||||
|
} catch (InterruptedException e) {
|
||||||
|
log.error("Interrupted while getting records from the cache", e);
|
||||||
|
}
|
||||||
|
return result;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public GetRecordsRetrievalStrategy getGetRecordsRetrievalStrategy() {
|
||||||
|
return getRecordsRetrievalStrategy;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void shutdown() {
|
||||||
|
defaultGetRecordsCacheDaemon.isShutdown = true;
|
||||||
|
executorService.shutdownNow();
|
||||||
|
started = false;
|
||||||
|
}
|
||||||
|
|
||||||
|
private class DefaultGetRecordsCacheDaemon implements Runnable {
|
||||||
|
volatile boolean isShutdown = false;
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void run() {
|
||||||
|
while (!isShutdown) {
|
||||||
|
if (Thread.currentThread().isInterrupted()) {
|
||||||
|
log.warn("Prefetch thread was interrupted.");
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
MetricsHelper.startScope(metricsFactory, operation);
|
||||||
|
if (prefetchCounters.shouldGetNewRecords()) {
|
||||||
|
try {
|
||||||
|
sleepBeforeNextCall();
|
||||||
|
GetRecordsResult getRecordsResult = getRecordsRetrievalStrategy.getRecords(maxRecordsPerCall);
|
||||||
|
lastSuccessfulCall = Instant.now();
|
||||||
|
ProcessRecordsInput processRecordsInput = new ProcessRecordsInput()
|
||||||
|
.withRecords(getRecordsResult.getRecords())
|
||||||
|
.withMillisBehindLatest(getRecordsResult.getMillisBehindLatest())
|
||||||
|
.withCacheEntryTime(lastSuccessfulCall);
|
||||||
|
getRecordsResultQueue.put(processRecordsInput);
|
||||||
|
prefetchCounters.added(processRecordsInput);
|
||||||
|
} catch (InterruptedException e) {
|
||||||
|
log.info("Thread was interrupted, indicating shutdown was called on the cache.");
|
||||||
|
} catch (SdkClientException e) {
|
||||||
|
log.error("Exception thrown while fetching records from Kinesis", e);
|
||||||
|
} catch (Throwable e) {
|
||||||
|
log.error("Unexpected exception was thrown. This could probably be an issue or a bug." +
|
||||||
|
" Please search for the exception/error online to check what is going on. If the " +
|
||||||
|
"issue persists or is a recurring problem, feel free to open an issue on, " +
|
||||||
|
"https://github.com/awslabs/amazon-kinesis-client.", e);
|
||||||
|
} finally {
|
||||||
|
MetricsHelper.endScope();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
callShutdownOnStrategy();
|
||||||
|
}
|
||||||
|
|
||||||
|
private void callShutdownOnStrategy() {
|
||||||
|
if (!getRecordsRetrievalStrategy.isShutdown()) {
|
||||||
|
getRecordsRetrievalStrategy.shutdown();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
private void sleepBeforeNextCall() throws InterruptedException {
|
||||||
|
if (lastSuccessfulCall == null) {
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
long timeSinceLastCall = Duration.between(lastSuccessfulCall, Instant.now()).abs().toMillis();
|
||||||
|
if (timeSinceLastCall < idleMillisBetweenCalls) {
|
||||||
|
Thread.sleep(idleMillisBetweenCalls - timeSinceLastCall);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
private class PrefetchCounters {
|
||||||
|
private long size = 0;
|
||||||
|
private long byteSize = 0;
|
||||||
|
|
||||||
|
public synchronized void added(final ProcessRecordsInput result) {
|
||||||
|
size += getSize(result);
|
||||||
|
byteSize += getByteSize(result);
|
||||||
|
}
|
||||||
|
|
||||||
|
public synchronized void removed(final ProcessRecordsInput result) {
|
||||||
|
size -= getSize(result);
|
||||||
|
byteSize -= getByteSize(result);
|
||||||
|
}
|
||||||
|
|
||||||
|
private long getSize(final ProcessRecordsInput result) {
|
||||||
|
return result.getRecords().size();
|
||||||
|
}
|
||||||
|
|
||||||
|
private long getByteSize(final ProcessRecordsInput result) {
|
||||||
|
return result.getRecords().stream().mapToLong(record -> record.getData().array().length).sum();
|
||||||
|
}
|
||||||
|
|
||||||
|
public synchronized boolean shouldGetNewRecords() {
|
||||||
|
return size < maxRecordsCount && byteSize < maxByteSize;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
@ -18,7 +18,6 @@ import java.math.BigInteger;
|
||||||
import java.util.Collections;
|
import java.util.Collections;
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
import java.util.ListIterator;
|
import java.util.ListIterator;
|
||||||
import java.util.Optional;
|
|
||||||
|
|
||||||
import org.apache.commons.logging.Log;
|
import org.apache.commons.logging.Log;
|
||||||
import org.apache.commons.logging.LogFactory;
|
import org.apache.commons.logging.LogFactory;
|
||||||
|
|
@ -63,7 +62,7 @@ class ProcessTask implements ITask {
|
||||||
private final Shard shard;
|
private final Shard shard;
|
||||||
private final ThrottlingReporter throttlingReporter;
|
private final ThrottlingReporter throttlingReporter;
|
||||||
|
|
||||||
private final GetRecordsRetrievalStrategy getRecordsRetrievalStrategy;
|
private final GetRecordsCache getRecordsCache;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* @param shardInfo
|
* @param shardInfo
|
||||||
|
|
@ -78,17 +77,17 @@ class ProcessTask implements ITask {
|
||||||
* Kinesis data fetcher (used to fetch records from Kinesis)
|
* Kinesis data fetcher (used to fetch records from Kinesis)
|
||||||
* @param backoffTimeMillis
|
* @param backoffTimeMillis
|
||||||
* backoff time when catching exceptions
|
* backoff time when catching exceptions
|
||||||
* @param getRecordsRetrievalStrategy
|
* @param getRecordsCache
|
||||||
* The retrieval strategy for fetching records from kinesis
|
* The retrieval strategy for fetching records from kinesis
|
||||||
*/
|
*/
|
||||||
public ProcessTask(ShardInfo shardInfo, StreamConfig streamConfig, IRecordProcessor recordProcessor,
|
public ProcessTask(ShardInfo shardInfo, StreamConfig streamConfig, IRecordProcessor recordProcessor,
|
||||||
RecordProcessorCheckpointer recordProcessorCheckpointer, KinesisDataFetcher dataFetcher,
|
RecordProcessorCheckpointer recordProcessorCheckpointer, KinesisDataFetcher dataFetcher,
|
||||||
long backoffTimeMillis, boolean skipShardSyncAtWorkerInitializationIfLeasesExist,
|
long backoffTimeMillis, boolean skipShardSyncAtWorkerInitializationIfLeasesExist,
|
||||||
GetRecordsRetrievalStrategy getRecordsRetrievalStrategy) {
|
GetRecordsCache getRecordsCache) {
|
||||||
this(shardInfo, streamConfig, recordProcessor, recordProcessorCheckpointer, dataFetcher, backoffTimeMillis,
|
this(shardInfo, streamConfig, recordProcessor, recordProcessorCheckpointer, dataFetcher, backoffTimeMillis,
|
||||||
skipShardSyncAtWorkerInitializationIfLeasesExist,
|
skipShardSyncAtWorkerInitializationIfLeasesExist,
|
||||||
new ThrottlingReporter(MAX_CONSECUTIVE_THROTTLES, shardInfo.getShardId()),
|
new ThrottlingReporter(MAX_CONSECUTIVE_THROTTLES, shardInfo.getShardId()),
|
||||||
getRecordsRetrievalStrategy);
|
getRecordsCache);
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
|
@ -108,9 +107,9 @@ class ProcessTask implements ITask {
|
||||||
* determines how throttling events should be reported in the log.
|
* determines how throttling events should be reported in the log.
|
||||||
*/
|
*/
|
||||||
public ProcessTask(ShardInfo shardInfo, StreamConfig streamConfig, IRecordProcessor recordProcessor,
|
public ProcessTask(ShardInfo shardInfo, StreamConfig streamConfig, IRecordProcessor recordProcessor,
|
||||||
RecordProcessorCheckpointer recordProcessorCheckpointer, KinesisDataFetcher dataFetcher,
|
RecordProcessorCheckpointer recordProcessorCheckpointer, KinesisDataFetcher dataFetcher,
|
||||||
long backoffTimeMillis, boolean skipShardSyncAtWorkerInitializationIfLeasesExist,
|
long backoffTimeMillis, boolean skipShardSyncAtWorkerInitializationIfLeasesExist,
|
||||||
ThrottlingReporter throttlingReporter, GetRecordsRetrievalStrategy getRecordsRetrievalStrategy) {
|
ThrottlingReporter throttlingReporter, GetRecordsCache getRecordsCache) {
|
||||||
super();
|
super();
|
||||||
this.shardInfo = shardInfo;
|
this.shardInfo = shardInfo;
|
||||||
this.recordProcessor = recordProcessor;
|
this.recordProcessor = recordProcessor;
|
||||||
|
|
@ -120,7 +119,7 @@ class ProcessTask implements ITask {
|
||||||
this.backoffTimeMillis = backoffTimeMillis;
|
this.backoffTimeMillis = backoffTimeMillis;
|
||||||
this.throttlingReporter = throttlingReporter;
|
this.throttlingReporter = throttlingReporter;
|
||||||
IKinesisProxy kinesisProxy = this.streamConfig.getStreamProxy();
|
IKinesisProxy kinesisProxy = this.streamConfig.getStreamProxy();
|
||||||
this.getRecordsRetrievalStrategy = getRecordsRetrievalStrategy;
|
this.getRecordsCache = getRecordsCache;
|
||||||
// If skipShardSyncAtWorkerInitializationIfLeasesExist is set, we will not get the shard for
|
// If skipShardSyncAtWorkerInitializationIfLeasesExist is set, we will not get the shard for
|
||||||
// this ProcessTask. In this case, duplicate KPL user records in the event of resharding will
|
// this ProcessTask. In this case, duplicate KPL user records in the event of resharding will
|
||||||
// not be dropped during deaggregation of Amazon Kinesis records. This is only applicable if
|
// not be dropped during deaggregation of Amazon Kinesis records. This is only applicable if
|
||||||
|
|
@ -149,7 +148,6 @@ class ProcessTask implements ITask {
|
||||||
scope.addDimension(MetricsHelper.SHARD_ID_DIMENSION_NAME, shardInfo.getShardId());
|
scope.addDimension(MetricsHelper.SHARD_ID_DIMENSION_NAME, shardInfo.getShardId());
|
||||||
scope.addData(RECORDS_PROCESSED_METRIC, 0, StandardUnit.Count, MetricsLevel.SUMMARY);
|
scope.addData(RECORDS_PROCESSED_METRIC, 0, StandardUnit.Count, MetricsLevel.SUMMARY);
|
||||||
scope.addData(DATA_BYTES_PROCESSED_METRIC, 0, StandardUnit.Bytes, MetricsLevel.SUMMARY);
|
scope.addData(DATA_BYTES_PROCESSED_METRIC, 0, StandardUnit.Bytes, MetricsLevel.SUMMARY);
|
||||||
|
|
||||||
Exception exception = null;
|
Exception exception = null;
|
||||||
|
|
||||||
try {
|
try {
|
||||||
|
|
@ -158,9 +156,9 @@ class ProcessTask implements ITask {
|
||||||
return new TaskResult(null, true);
|
return new TaskResult(null, true);
|
||||||
}
|
}
|
||||||
|
|
||||||
final GetRecordsResult getRecordsResult = getRecordsResult();
|
final ProcessRecordsInput processRecordsInput = getRecordsResult();
|
||||||
throttlingReporter.success();
|
throttlingReporter.success();
|
||||||
List<Record> records = getRecordsResult.getRecords();
|
List<Record> records = processRecordsInput.getRecords();
|
||||||
|
|
||||||
if (!records.isEmpty()) {
|
if (!records.isEmpty()) {
|
||||||
scope.addData(RECORDS_PROCESSED_METRIC, records.size(), StandardUnit.Count, MetricsLevel.SUMMARY);
|
scope.addData(RECORDS_PROCESSED_METRIC, records.size(), StandardUnit.Count, MetricsLevel.SUMMARY);
|
||||||
|
|
@ -175,7 +173,7 @@ class ProcessTask implements ITask {
|
||||||
recordProcessorCheckpointer.getLargestPermittedCheckpointValue()));
|
recordProcessorCheckpointer.getLargestPermittedCheckpointValue()));
|
||||||
|
|
||||||
if (shouldCallProcessRecords(records)) {
|
if (shouldCallProcessRecords(records)) {
|
||||||
callProcessRecords(getRecordsResult, records);
|
callProcessRecords(processRecordsInput, records);
|
||||||
}
|
}
|
||||||
} catch (ProvisionedThroughputExceededException pte) {
|
} catch (ProvisionedThroughputExceededException pte) {
|
||||||
throttlingReporter.throttled();
|
throttlingReporter.throttled();
|
||||||
|
|
@ -206,17 +204,17 @@ class ProcessTask implements ITask {
|
||||||
/**
|
/**
|
||||||
* Dispatches a batch of records to the record processor, and handles any fallout from that.
|
* Dispatches a batch of records to the record processor, and handles any fallout from that.
|
||||||
*
|
*
|
||||||
* @param getRecordsResult
|
* @param input
|
||||||
* the result of the last call to Kinesis
|
* the result of the last call to Kinesis
|
||||||
* @param records
|
* @param records
|
||||||
* the records to be dispatched. It's possible the records have been adjusted by KPL deaggregation.
|
* the records to be dispatched. It's possible the records have been adjusted by KPL deaggregation.
|
||||||
*/
|
*/
|
||||||
private void callProcessRecords(GetRecordsResult getRecordsResult, List<Record> records) {
|
private void callProcessRecords(ProcessRecordsInput input, List<Record> records) {
|
||||||
LOG.debug("Calling application processRecords() with " + records.size() + " records from "
|
LOG.debug("Calling application processRecords() with " + records.size() + " records from "
|
||||||
+ shardInfo.getShardId());
|
+ shardInfo.getShardId());
|
||||||
final ProcessRecordsInput processRecordsInput = new ProcessRecordsInput().withRecords(records)
|
final ProcessRecordsInput processRecordsInput = new ProcessRecordsInput().withRecords(records)
|
||||||
.withCheckpointer(recordProcessorCheckpointer)
|
.withCheckpointer(recordProcessorCheckpointer)
|
||||||
.withMillisBehindLatest(getRecordsResult.getMillisBehindLatest());
|
.withMillisBehindLatest(input.getMillisBehindLatest());
|
||||||
|
|
||||||
final long recordProcessorStartTimeMillis = System.currentTimeMillis();
|
final long recordProcessorStartTimeMillis = System.currentTimeMillis();
|
||||||
try {
|
try {
|
||||||
|
|
@ -304,8 +302,8 @@ class ProcessTask implements ITask {
|
||||||
* @return the largest extended sequence number among the retained records
|
* @return the largest extended sequence number among the retained records
|
||||||
*/
|
*/
|
||||||
private ExtendedSequenceNumber filterAndGetMaxExtendedSequenceNumber(IMetricsScope scope, List<Record> records,
|
private ExtendedSequenceNumber filterAndGetMaxExtendedSequenceNumber(IMetricsScope scope, List<Record> records,
|
||||||
final ExtendedSequenceNumber lastCheckpointValue,
|
final ExtendedSequenceNumber lastCheckpointValue,
|
||||||
final ExtendedSequenceNumber lastLargestPermittedCheckpointValue) {
|
final ExtendedSequenceNumber lastLargestPermittedCheckpointValue) {
|
||||||
ExtendedSequenceNumber largestExtendedSequenceNumber = lastLargestPermittedCheckpointValue;
|
ExtendedSequenceNumber largestExtendedSequenceNumber = lastLargestPermittedCheckpointValue;
|
||||||
ListIterator<Record> recordIterator = records.listIterator();
|
ListIterator<Record> recordIterator = records.listIterator();
|
||||||
while (recordIterator.hasNext()) {
|
while (recordIterator.hasNext()) {
|
||||||
|
|
@ -339,7 +337,7 @@ class ProcessTask implements ITask {
|
||||||
*
|
*
|
||||||
* @return list of data records from Kinesis
|
* @return list of data records from Kinesis
|
||||||
*/
|
*/
|
||||||
private GetRecordsResult getRecordsResult() {
|
private ProcessRecordsInput getRecordsResult() {
|
||||||
try {
|
try {
|
||||||
return getRecordsResultAndRecordMillisBehindLatest();
|
return getRecordsResultAndRecordMillisBehindLatest();
|
||||||
} catch (ExpiredIteratorException e) {
|
} catch (ExpiredIteratorException e) {
|
||||||
|
|
@ -375,22 +373,17 @@ class ProcessTask implements ITask {
|
||||||
*
|
*
|
||||||
* @return list of data records from Kinesis
|
* @return list of data records from Kinesis
|
||||||
*/
|
*/
|
||||||
private GetRecordsResult getRecordsResultAndRecordMillisBehindLatest() {
|
private ProcessRecordsInput getRecordsResultAndRecordMillisBehindLatest() {
|
||||||
final GetRecordsResult getRecordsResult = getRecordsRetrievalStrategy.getRecords(streamConfig.getMaxRecords());
|
final ProcessRecordsInput processRecordsInput = getRecordsCache.getNextResult();
|
||||||
|
|
||||||
if (getRecordsResult == null) {
|
if (processRecordsInput.getMillisBehindLatest() != null) {
|
||||||
// Stream no longer exists
|
|
||||||
return new GetRecordsResult().withRecords(Collections.<Record>emptyList());
|
|
||||||
}
|
|
||||||
|
|
||||||
if (getRecordsResult.getMillisBehindLatest() != null) {
|
|
||||||
MetricsHelper.getMetricsScope().addData(MILLIS_BEHIND_LATEST_METRIC,
|
MetricsHelper.getMetricsScope().addData(MILLIS_BEHIND_LATEST_METRIC,
|
||||||
getRecordsResult.getMillisBehindLatest(),
|
processRecordsInput.getMillisBehindLatest(),
|
||||||
StandardUnit.Milliseconds,
|
StandardUnit.Milliseconds,
|
||||||
MetricsLevel.SUMMARY);
|
MetricsLevel.SUMMARY);
|
||||||
}
|
}
|
||||||
|
|
||||||
return getRecordsResult;
|
return processRecordsInput;
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
@ -0,0 +1,74 @@
|
||||||
|
/*
|
||||||
|
* Copyright 2017 Amazon.com, Inc. or its affiliates. All Rights Reserved.
|
||||||
|
*
|
||||||
|
* Licensed under the Amazon Software License (the "License").
|
||||||
|
* You may not use this file except in compliance with the License.
|
||||||
|
* A copy of the License is located at
|
||||||
|
*
|
||||||
|
* http://aws.amazon.com/asl/
|
||||||
|
*
|
||||||
|
* or in the "license" file accompanying this file. This file is distributed
|
||||||
|
* on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
|
||||||
|
* express or implied. See the License for the specific language governing
|
||||||
|
* permissions and limitations under the License.
|
||||||
|
*/
|
||||||
|
package com.amazonaws.services.kinesis.clientlibrary.lib.worker;
|
||||||
|
|
||||||
|
import com.amazonaws.services.kinesis.metrics.interfaces.IMetricsFactory;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* This factory is used to create the records fetcher to retrieve data from Kinesis for a given shard.
|
||||||
|
*/
|
||||||
|
public interface RecordsFetcherFactory {
|
||||||
|
/**
|
||||||
|
* Returns a GetRecordsCache to be used for retrieving records for a given shard.
|
||||||
|
*
|
||||||
|
* @param getRecordsRetrievalStrategy GetRecordsRetrievalStrategy to be used with the GetRecordsCache
|
||||||
|
* @param shardId ShardId of the shard that the fetcher will retrieve records for
|
||||||
|
* @param metricsFactory MetricsFactory used to create metricScope
|
||||||
|
*
|
||||||
|
* @return GetRecordsCache used to get records from Kinesis.
|
||||||
|
*/
|
||||||
|
GetRecordsCache createRecordsFetcher(GetRecordsRetrievalStrategy getRecordsRetrievalStrategy, String shardId, IMetricsFactory metricsFactory);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Sets the maximum number of ProcessRecordsInput objects the GetRecordsCache can hold, before further requests are
|
||||||
|
* blocked.
|
||||||
|
*
|
||||||
|
* @param maxPendingProcessRecordsInput The maximum number of ProcessRecordsInput objects that the cache will accept
|
||||||
|
* before blocking.
|
||||||
|
*/
|
||||||
|
void setMaxPendingProcessRecordsInput(int maxPendingProcessRecordsInput);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Sets the max byte size for the GetRecordsCache, before further requests are blocked. The byte size of the cache
|
||||||
|
* is the sum of byte size of all the ProcessRecordsInput objects in the cache at any point of time.
|
||||||
|
*
|
||||||
|
* @param maxByteSize The maximum byte size for the cache before blocking.
|
||||||
|
*/
|
||||||
|
void setMaxByteSize(int maxByteSize);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Sets the max number of records for the GetRecordsCache can hold, before further requests are blocked. The records
|
||||||
|
* count is the sum of all records present in across all the ProcessRecordsInput objects in the cache at any point
|
||||||
|
* of time.
|
||||||
|
*
|
||||||
|
* @param maxRecordsCount The mximum number of records in the cache before blocking.
|
||||||
|
*/
|
||||||
|
void setMaxRecordsCount(int maxRecordsCount);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Sets the dataFetchingStrategy to determine the type of GetRecordsCache to be used.
|
||||||
|
*
|
||||||
|
* @param dataFetchingStrategy Fetching strategy to be used
|
||||||
|
*/
|
||||||
|
void setDataFetchingStrategy(DataFetchingStrategy dataFetchingStrategy);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Sets the maximum idle time between two get calls.
|
||||||
|
*
|
||||||
|
* @param idleMillisBetweenCalls Sleep millis between calls.
|
||||||
|
*/
|
||||||
|
void setIdleMillisBetweenCalls(long idleMillisBetweenCalls);
|
||||||
|
|
||||||
|
}
|
||||||
|
|
@ -20,7 +20,6 @@ import java.util.concurrent.ExecutorService;
|
||||||
import java.util.concurrent.Future;
|
import java.util.concurrent.Future;
|
||||||
import java.util.concurrent.RejectedExecutionException;
|
import java.util.concurrent.RejectedExecutionException;
|
||||||
|
|
||||||
import lombok.Getter;
|
|
||||||
import org.apache.commons.logging.Log;
|
import org.apache.commons.logging.Log;
|
||||||
import org.apache.commons.logging.LogFactory;
|
import org.apache.commons.logging.LogFactory;
|
||||||
|
|
||||||
|
|
@ -32,6 +31,8 @@ import com.amazonaws.services.kinesis.leases.interfaces.ILeaseManager;
|
||||||
import com.amazonaws.services.kinesis.metrics.interfaces.IMetricsFactory;
|
import com.amazonaws.services.kinesis.metrics.interfaces.IMetricsFactory;
|
||||||
import com.google.common.annotations.VisibleForTesting;
|
import com.google.common.annotations.VisibleForTesting;
|
||||||
|
|
||||||
|
import lombok.Getter;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Responsible for consuming data records of a (specified) shard.
|
* Responsible for consuming data records of a (specified) shard.
|
||||||
* The instance should be shutdown when we lose the primary responsibility for a shard.
|
* The instance should be shutdown when we lose the primary responsibility for a shard.
|
||||||
|
|
@ -43,6 +44,7 @@ class ShardConsumer {
|
||||||
|
|
||||||
private final StreamConfig streamConfig;
|
private final StreamConfig streamConfig;
|
||||||
private final IRecordProcessor recordProcessor;
|
private final IRecordProcessor recordProcessor;
|
||||||
|
private final KinesisClientLibConfiguration config;
|
||||||
private final RecordProcessorCheckpointer recordProcessorCheckpointer;
|
private final RecordProcessorCheckpointer recordProcessorCheckpointer;
|
||||||
private final ExecutorService executorService;
|
private final ExecutorService executorService;
|
||||||
private final ShardInfo shardInfo;
|
private final ShardInfo shardInfo;
|
||||||
|
|
@ -61,7 +63,7 @@ class ShardConsumer {
|
||||||
private Future<TaskResult> future;
|
private Future<TaskResult> future;
|
||||||
|
|
||||||
@Getter
|
@Getter
|
||||||
private final GetRecordsRetrievalStrategy getRecordsRetrievalStrategy;
|
private final GetRecordsCache getRecordsCache;
|
||||||
|
|
||||||
private static final GetRecordsRetrievalStrategy makeStrategy(KinesisDataFetcher dataFetcher,
|
private static final GetRecordsRetrievalStrategy makeStrategy(KinesisDataFetcher dataFetcher,
|
||||||
Optional<Integer> retryGetRecordsInSeconds,
|
Optional<Integer> retryGetRecordsInSeconds,
|
||||||
|
|
@ -91,6 +93,7 @@ class ShardConsumer {
|
||||||
* @param streamConfig Stream configuration to use
|
* @param streamConfig Stream configuration to use
|
||||||
* @param checkpoint Checkpoint tracker
|
* @param checkpoint Checkpoint tracker
|
||||||
* @param recordProcessor Record processor used to process the data records for the shard
|
* @param recordProcessor Record processor used to process the data records for the shard
|
||||||
|
* @param config Kinesis library configuration
|
||||||
* @param leaseManager Used to create leases for new shards
|
* @param leaseManager Used to create leases for new shards
|
||||||
* @param parentShardPollIntervalMillis Wait for this long if parent shards are not done (or we get an exception)
|
* @param parentShardPollIntervalMillis Wait for this long if parent shards are not done (or we get an exception)
|
||||||
* @param executorService ExecutorService used to execute process tasks for this shard
|
* @param executorService ExecutorService used to execute process tasks for this shard
|
||||||
|
|
@ -99,19 +102,31 @@ class ShardConsumer {
|
||||||
*/
|
*/
|
||||||
// CHECKSTYLE:IGNORE ParameterNumber FOR NEXT 10 LINES
|
// CHECKSTYLE:IGNORE ParameterNumber FOR NEXT 10 LINES
|
||||||
ShardConsumer(ShardInfo shardInfo,
|
ShardConsumer(ShardInfo shardInfo,
|
||||||
StreamConfig streamConfig,
|
StreamConfig streamConfig,
|
||||||
ICheckpoint checkpoint,
|
ICheckpoint checkpoint,
|
||||||
IRecordProcessor recordProcessor,
|
IRecordProcessor recordProcessor,
|
||||||
ILeaseManager<KinesisClientLease> leaseManager,
|
ILeaseManager<KinesisClientLease> leaseManager,
|
||||||
long parentShardPollIntervalMillis,
|
long parentShardPollIntervalMillis,
|
||||||
boolean cleanupLeasesOfCompletedShards,
|
boolean cleanupLeasesOfCompletedShards,
|
||||||
ExecutorService executorService,
|
ExecutorService executorService,
|
||||||
IMetricsFactory metricsFactory,
|
IMetricsFactory metricsFactory,
|
||||||
long backoffTimeMillis,
|
long backoffTimeMillis,
|
||||||
boolean skipShardSyncAtWorkerInitializationIfLeasesExist) {
|
boolean skipShardSyncAtWorkerInitializationIfLeasesExist,
|
||||||
this(shardInfo, streamConfig, checkpoint,recordProcessor, leaseManager, parentShardPollIntervalMillis,
|
KinesisClientLibConfiguration config) {
|
||||||
cleanupLeasesOfCompletedShards, executorService, metricsFactory, backoffTimeMillis,
|
this(shardInfo,
|
||||||
skipShardSyncAtWorkerInitializationIfLeasesExist, Optional.empty(), Optional.empty());
|
streamConfig,
|
||||||
|
checkpoint,
|
||||||
|
recordProcessor,
|
||||||
|
leaseManager,
|
||||||
|
parentShardPollIntervalMillis,
|
||||||
|
cleanupLeasesOfCompletedShards,
|
||||||
|
executorService,
|
||||||
|
metricsFactory,
|
||||||
|
backoffTimeMillis,
|
||||||
|
skipShardSyncAtWorkerInitializationIfLeasesExist,
|
||||||
|
Optional.empty(),
|
||||||
|
Optional.empty(),
|
||||||
|
config);
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
|
@ -126,6 +141,7 @@ class ShardConsumer {
|
||||||
* @param backoffTimeMillis backoff interval when we encounter exceptions
|
* @param backoffTimeMillis backoff interval when we encounter exceptions
|
||||||
* @param retryGetRecordsInSeconds time in seconds to wait before the worker retries to get a record.
|
* @param retryGetRecordsInSeconds time in seconds to wait before the worker retries to get a record.
|
||||||
* @param maxGetRecordsThreadPool max number of threads in the getRecords thread pool.
|
* @param maxGetRecordsThreadPool max number of threads in the getRecords thread pool.
|
||||||
|
* @param config Kinesis library configuration
|
||||||
*/
|
*/
|
||||||
// CHECKSTYLE:IGNORE ParameterNumber FOR NEXT 10 LINES
|
// CHECKSTYLE:IGNORE ParameterNumber FOR NEXT 10 LINES
|
||||||
ShardConsumer(ShardInfo shardInfo,
|
ShardConsumer(ShardInfo shardInfo,
|
||||||
|
|
@ -140,26 +156,86 @@ class ShardConsumer {
|
||||||
long backoffTimeMillis,
|
long backoffTimeMillis,
|
||||||
boolean skipShardSyncAtWorkerInitializationIfLeasesExist,
|
boolean skipShardSyncAtWorkerInitializationIfLeasesExist,
|
||||||
Optional<Integer> retryGetRecordsInSeconds,
|
Optional<Integer> retryGetRecordsInSeconds,
|
||||||
Optional<Integer> maxGetRecordsThreadPool) {
|
Optional<Integer> maxGetRecordsThreadPool,
|
||||||
this.streamConfig = streamConfig;
|
KinesisClientLibConfiguration config) {
|
||||||
this.recordProcessor = recordProcessor;
|
|
||||||
this.executorService = executorService;
|
this(
|
||||||
this.shardInfo = shardInfo;
|
shardInfo,
|
||||||
this.checkpoint = checkpoint;
|
streamConfig,
|
||||||
this.recordProcessorCheckpointer =
|
checkpoint,
|
||||||
new RecordProcessorCheckpointer(shardInfo,
|
recordProcessor,
|
||||||
|
new RecordProcessorCheckpointer(
|
||||||
|
shardInfo,
|
||||||
checkpoint,
|
checkpoint,
|
||||||
new SequenceNumberValidator(streamConfig.getStreamProxy(),
|
new SequenceNumberValidator(
|
||||||
|
streamConfig.getStreamProxy(),
|
||||||
shardInfo.getShardId(),
|
shardInfo.getShardId(),
|
||||||
streamConfig.shouldValidateSequenceNumberBeforeCheckpointing()));
|
streamConfig.shouldValidateSequenceNumberBeforeCheckpointing())),
|
||||||
this.dataFetcher = new KinesisDataFetcher(streamConfig.getStreamProxy(), shardInfo);
|
leaseManager,
|
||||||
|
parentShardPollIntervalMillis,
|
||||||
|
cleanupLeasesOfCompletedShards,
|
||||||
|
executorService,
|
||||||
|
metricsFactory,
|
||||||
|
backoffTimeMillis,
|
||||||
|
skipShardSyncAtWorkerInitializationIfLeasesExist,
|
||||||
|
new KinesisDataFetcher(streamConfig.getStreamProxy(), shardInfo),
|
||||||
|
retryGetRecordsInSeconds,
|
||||||
|
maxGetRecordsThreadPool,
|
||||||
|
config
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @param shardInfo Shard information
|
||||||
|
* @param streamConfig Stream Config to use
|
||||||
|
* @param checkpoint Checkpoint tracker
|
||||||
|
* @param recordProcessor Record processor used to process the data records for the shard
|
||||||
|
* @param recordProcessorCheckpointer RecordProcessorCheckpointer to use to checkpoint progress
|
||||||
|
* @param leaseManager Used to create leases for new shards
|
||||||
|
* @param parentShardPollIntervalMillis Wait for this long if parent shards are not done (or we get an exception)
|
||||||
|
* @param cleanupLeasesOfCompletedShards clean up the leases of completed shards
|
||||||
|
* @param executorService ExecutorService used to execute process tasks for this shard
|
||||||
|
* @param metricsFactory IMetricsFactory used to construct IMetricsScopes for this shard
|
||||||
|
* @param backoffTimeMillis backoff interval when we encounter exceptions
|
||||||
|
* @param skipShardSyncAtWorkerInitializationIfLeasesExist Skip sync at init if lease exists
|
||||||
|
* @param kinesisDataFetcher KinesisDataFetcher to fetch data from Kinesis streams.
|
||||||
|
* @param retryGetRecordsInSeconds time in seconds to wait before the worker retries to get a record
|
||||||
|
* @param maxGetRecordsThreadPool max number of threads in the getRecords thread pool
|
||||||
|
* @param config Kinesis library configuration
|
||||||
|
*/
|
||||||
|
ShardConsumer(ShardInfo shardInfo,
|
||||||
|
StreamConfig streamConfig,
|
||||||
|
ICheckpoint checkpoint,
|
||||||
|
IRecordProcessor recordProcessor,
|
||||||
|
RecordProcessorCheckpointer recordProcessorCheckpointer,
|
||||||
|
ILeaseManager<KinesisClientLease> leaseManager,
|
||||||
|
long parentShardPollIntervalMillis,
|
||||||
|
boolean cleanupLeasesOfCompletedShards,
|
||||||
|
ExecutorService executorService,
|
||||||
|
IMetricsFactory metricsFactory,
|
||||||
|
long backoffTimeMillis,
|
||||||
|
boolean skipShardSyncAtWorkerInitializationIfLeasesExist,
|
||||||
|
KinesisDataFetcher kinesisDataFetcher,
|
||||||
|
Optional<Integer> retryGetRecordsInSeconds,
|
||||||
|
Optional<Integer> maxGetRecordsThreadPool,
|
||||||
|
KinesisClientLibConfiguration config) {
|
||||||
|
this.shardInfo = shardInfo;
|
||||||
|
this.streamConfig = streamConfig;
|
||||||
|
this.checkpoint = checkpoint;
|
||||||
|
this.recordProcessor = recordProcessor;
|
||||||
|
this.recordProcessorCheckpointer = recordProcessorCheckpointer;
|
||||||
this.leaseManager = leaseManager;
|
this.leaseManager = leaseManager;
|
||||||
this.metricsFactory = metricsFactory;
|
|
||||||
this.parentShardPollIntervalMillis = parentShardPollIntervalMillis;
|
this.parentShardPollIntervalMillis = parentShardPollIntervalMillis;
|
||||||
this.cleanupLeasesOfCompletedShards = cleanupLeasesOfCompletedShards;
|
this.cleanupLeasesOfCompletedShards = cleanupLeasesOfCompletedShards;
|
||||||
|
this.executorService = executorService;
|
||||||
|
this.metricsFactory = metricsFactory;
|
||||||
this.taskBackoffTimeMillis = backoffTimeMillis;
|
this.taskBackoffTimeMillis = backoffTimeMillis;
|
||||||
this.skipShardSyncAtWorkerInitializationIfLeasesExist = skipShardSyncAtWorkerInitializationIfLeasesExist;
|
this.skipShardSyncAtWorkerInitializationIfLeasesExist = skipShardSyncAtWorkerInitializationIfLeasesExist;
|
||||||
this.getRecordsRetrievalStrategy = makeStrategy(dataFetcher, retryGetRecordsInSeconds, maxGetRecordsThreadPool, shardInfo);
|
this.config = config;
|
||||||
|
this.dataFetcher = kinesisDataFetcher;
|
||||||
|
this.getRecordsCache = config.getRecordsFetcherFactory().createRecordsFetcher(
|
||||||
|
makeStrategy(this.dataFetcher, retryGetRecordsInSeconds, maxGetRecordsThreadPool, this.shardInfo),
|
||||||
|
this.getShardInfo().getShardId(), this.metricsFactory);
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
|
|
||||||
|
|
@ -46,7 +46,7 @@ class ShutdownTask implements ITask {
|
||||||
private final boolean cleanupLeasesOfCompletedShards;
|
private final boolean cleanupLeasesOfCompletedShards;
|
||||||
private final TaskType taskType = TaskType.SHUTDOWN;
|
private final TaskType taskType = TaskType.SHUTDOWN;
|
||||||
private final long backoffTimeMillis;
|
private final long backoffTimeMillis;
|
||||||
private final GetRecordsRetrievalStrategy getRecordsRetrievalStrategy;
|
private final GetRecordsCache getRecordsCache;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Constructor.
|
* Constructor.
|
||||||
|
|
@ -61,7 +61,7 @@ class ShutdownTask implements ITask {
|
||||||
boolean cleanupLeasesOfCompletedShards,
|
boolean cleanupLeasesOfCompletedShards,
|
||||||
ILeaseManager<KinesisClientLease> leaseManager,
|
ILeaseManager<KinesisClientLease> leaseManager,
|
||||||
long backoffTimeMillis,
|
long backoffTimeMillis,
|
||||||
GetRecordsRetrievalStrategy getRecordsRetrievalStrategy) {
|
GetRecordsCache getRecordsCache) {
|
||||||
this.shardInfo = shardInfo;
|
this.shardInfo = shardInfo;
|
||||||
this.recordProcessor = recordProcessor;
|
this.recordProcessor = recordProcessor;
|
||||||
this.recordProcessorCheckpointer = recordProcessorCheckpointer;
|
this.recordProcessorCheckpointer = recordProcessorCheckpointer;
|
||||||
|
|
@ -71,7 +71,7 @@ class ShutdownTask implements ITask {
|
||||||
this.cleanupLeasesOfCompletedShards = cleanupLeasesOfCompletedShards;
|
this.cleanupLeasesOfCompletedShards = cleanupLeasesOfCompletedShards;
|
||||||
this.leaseManager = leaseManager;
|
this.leaseManager = leaseManager;
|
||||||
this.backoffTimeMillis = backoffTimeMillis;
|
this.backoffTimeMillis = backoffTimeMillis;
|
||||||
this.getRecordsRetrievalStrategy = getRecordsRetrievalStrategy;
|
this.getRecordsCache = getRecordsCache;
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
|
@ -111,7 +111,7 @@ class ShutdownTask implements ITask {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
LOG.debug("Shutting down retrieval strategy.");
|
LOG.debug("Shutting down retrieval strategy.");
|
||||||
getRecordsRetrievalStrategy.shutdown();
|
getRecordsCache.shutdown();
|
||||||
LOG.debug("Record processor completed shutdown() for shard " + shardInfo.getShardId());
|
LOG.debug("Record processor completed shutdown() for shard " + shardInfo.getShardId());
|
||||||
} catch (Exception e) {
|
} catch (Exception e) {
|
||||||
applicationException = true;
|
applicationException = true;
|
||||||
|
|
|
||||||
|
|
@ -0,0 +1,77 @@
|
||||||
|
/*
|
||||||
|
* Copyright 2017 Amazon.com, Inc. or its affiliates. All Rights Reserved.
|
||||||
|
*
|
||||||
|
* Licensed under the Amazon Software License (the "License").
|
||||||
|
* You may not use this file except in compliance with the License.
|
||||||
|
* A copy of the License is located at
|
||||||
|
*
|
||||||
|
* http://aws.amazon.com/asl/
|
||||||
|
*
|
||||||
|
* or in the "license" file accompanying this file. This file is distributed
|
||||||
|
* on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
|
||||||
|
* express or implied. See the License for the specific language governing
|
||||||
|
* permissions and limitations under the License.
|
||||||
|
*/
|
||||||
|
package com.amazonaws.services.kinesis.clientlibrary.lib.worker;
|
||||||
|
|
||||||
|
import java.util.concurrent.Executors;
|
||||||
|
|
||||||
|
import com.amazonaws.services.kinesis.metrics.interfaces.IMetricsFactory;
|
||||||
|
import com.google.common.util.concurrent.ThreadFactoryBuilder;
|
||||||
|
import lombok.extern.apachecommons.CommonsLog;
|
||||||
|
|
||||||
|
@CommonsLog
|
||||||
|
public class SimpleRecordsFetcherFactory implements RecordsFetcherFactory {
|
||||||
|
private final int maxRecords;
|
||||||
|
private int maxPendingProcessRecordsInput = 3;
|
||||||
|
private int maxByteSize = 8 * 1024 * 1024;
|
||||||
|
private int maxRecordsCount = 30000;
|
||||||
|
private long idleMillisBetweenCalls = 1500L;
|
||||||
|
private DataFetchingStrategy dataFetchingStrategy = DataFetchingStrategy.DEFAULT;
|
||||||
|
private IMetricsFactory metricsFactory;
|
||||||
|
|
||||||
|
public SimpleRecordsFetcherFactory(int maxRecords) {
|
||||||
|
this.maxRecords = maxRecords;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public GetRecordsCache createRecordsFetcher(GetRecordsRetrievalStrategy getRecordsRetrievalStrategy, String shardId, IMetricsFactory metricsFactory) {
|
||||||
|
if(dataFetchingStrategy.equals(DataFetchingStrategy.DEFAULT)) {
|
||||||
|
return new BlockingGetRecordsCache(maxRecords, getRecordsRetrievalStrategy, idleMillisBetweenCalls);
|
||||||
|
} else {
|
||||||
|
return new PrefetchGetRecordsCache(maxPendingProcessRecordsInput, maxByteSize, maxRecordsCount, maxRecords,
|
||||||
|
getRecordsRetrievalStrategy,
|
||||||
|
Executors.newFixedThreadPool(1, new ThreadFactoryBuilder()
|
||||||
|
.setDaemon(true)
|
||||||
|
.setNameFormat("prefetch-cache-" + shardId + "-%04d")
|
||||||
|
.build()),
|
||||||
|
idleMillisBetweenCalls,
|
||||||
|
metricsFactory,
|
||||||
|
"ProcessTask");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void setMaxPendingProcessRecordsInput(int maxPendingProcessRecordsInput){
|
||||||
|
this.maxPendingProcessRecordsInput = maxPendingProcessRecordsInput;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void setMaxByteSize(int maxByteSize){
|
||||||
|
this.maxByteSize = maxByteSize;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void setMaxRecordsCount(int maxRecordsCount) {
|
||||||
|
this.maxRecordsCount = maxRecordsCount;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void setDataFetchingStrategy(DataFetchingStrategy dataFetchingStrategy){
|
||||||
|
this.dataFetchingStrategy = dataFetchingStrategy;
|
||||||
|
}
|
||||||
|
|
||||||
|
public void setIdleMillisBetweenCalls(final long idleMillisBetweenCalls) {
|
||||||
|
this.idleMillisBetweenCalls = idleMillisBetweenCalls;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
@ -73,6 +73,7 @@ public class Worker implements Runnable {
|
||||||
|
|
||||||
private final String applicationName;
|
private final String applicationName;
|
||||||
private final IRecordProcessorFactory recordProcessorFactory;
|
private final IRecordProcessorFactory recordProcessorFactory;
|
||||||
|
private final KinesisClientLibConfiguration config;
|
||||||
private final StreamConfig streamConfig;
|
private final StreamConfig streamConfig;
|
||||||
private final InitialPositionInStreamExtended initialPosition;
|
private final InitialPositionInStreamExtended initialPosition;
|
||||||
private final ICheckpoint checkpointTracker;
|
private final ICheckpoint checkpointTracker;
|
||||||
|
|
@ -245,6 +246,7 @@ public class Worker implements Runnable {
|
||||||
KinesisClientLibConfiguration config, AmazonKinesis kinesisClient, AmazonDynamoDB dynamoDBClient,
|
KinesisClientLibConfiguration config, AmazonKinesis kinesisClient, AmazonDynamoDB dynamoDBClient,
|
||||||
IMetricsFactory metricsFactory, ExecutorService execService) {
|
IMetricsFactory metricsFactory, ExecutorService execService) {
|
||||||
this(config.getApplicationName(), new V1ToV2RecordProcessorFactoryAdapter(recordProcessorFactory),
|
this(config.getApplicationName(), new V1ToV2RecordProcessorFactoryAdapter(recordProcessorFactory),
|
||||||
|
config,
|
||||||
new StreamConfig(
|
new StreamConfig(
|
||||||
new KinesisProxyFactory(config.getKinesisCredentialsProvider(), kinesisClient)
|
new KinesisProxyFactory(config.getKinesisCredentialsProvider(), kinesisClient)
|
||||||
.getProxy(config.getStreamName()),
|
.getProxy(config.getStreamName()),
|
||||||
|
|
@ -306,6 +308,8 @@ public class Worker implements Runnable {
|
||||||
* Name of the Kinesis application
|
* Name of the Kinesis application
|
||||||
* @param recordProcessorFactory
|
* @param recordProcessorFactory
|
||||||
* Used to get record processor instances for processing data from shards
|
* Used to get record processor instances for processing data from shards
|
||||||
|
* @paran config
|
||||||
|
* Kinesis Library configuration
|
||||||
* @param streamConfig
|
* @param streamConfig
|
||||||
* Stream configuration
|
* Stream configuration
|
||||||
* @param initialPositionInStream
|
* @param initialPositionInStream
|
||||||
|
|
@ -333,24 +337,25 @@ public class Worker implements Runnable {
|
||||||
*/
|
*/
|
||||||
// NOTE: This has package level access solely for testing
|
// NOTE: This has package level access solely for testing
|
||||||
// CHECKSTYLE:IGNORE ParameterNumber FOR NEXT 10 LINES
|
// CHECKSTYLE:IGNORE ParameterNumber FOR NEXT 10 LINES
|
||||||
Worker(String applicationName, IRecordProcessorFactory recordProcessorFactory, StreamConfig streamConfig,
|
Worker(String applicationName, IRecordProcessorFactory recordProcessorFactory, KinesisClientLibConfiguration config,
|
||||||
InitialPositionInStreamExtended initialPositionInStream, long parentShardPollIntervalMillis,
|
StreamConfig streamConfig, InitialPositionInStreamExtended initialPositionInStream, long parentShardPollIntervalMillis,
|
||||||
long shardSyncIdleTimeMillis, boolean cleanupLeasesUponShardCompletion, ICheckpoint checkpoint,
|
long shardSyncIdleTimeMillis, boolean cleanupLeasesUponShardCompletion, ICheckpoint checkpoint,
|
||||||
KinesisClientLibLeaseCoordinator leaseCoordinator, ExecutorService execService,
|
KinesisClientLibLeaseCoordinator leaseCoordinator, ExecutorService execService,
|
||||||
IMetricsFactory metricsFactory, long taskBackoffTimeMillis, long failoverTimeMillis,
|
IMetricsFactory metricsFactory, long taskBackoffTimeMillis, long failoverTimeMillis,
|
||||||
boolean skipShardSyncAtWorkerInitializationIfLeasesExist, ShardPrioritization shardPrioritization) {
|
boolean skipShardSyncAtWorkerInitializationIfLeasesExist, ShardPrioritization shardPrioritization) {
|
||||||
this(applicationName, recordProcessorFactory, streamConfig, initialPositionInStream, parentShardPollIntervalMillis,
|
this(applicationName, recordProcessorFactory, config, streamConfig, initialPositionInStream, parentShardPollIntervalMillis,
|
||||||
shardSyncIdleTimeMillis, cleanupLeasesUponShardCompletion, checkpoint, leaseCoordinator, execService,
|
shardSyncIdleTimeMillis, cleanupLeasesUponShardCompletion, checkpoint, leaseCoordinator, execService,
|
||||||
metricsFactory, taskBackoffTimeMillis, failoverTimeMillis, skipShardSyncAtWorkerInitializationIfLeasesExist,
|
metricsFactory, taskBackoffTimeMillis, failoverTimeMillis, skipShardSyncAtWorkerInitializationIfLeasesExist,
|
||||||
shardPrioritization, Optional.empty(), Optional.empty());
|
shardPrioritization, Optional.empty(), Optional.empty());
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* @param applicationName
|
* @param applicationName
|
||||||
* Name of the Kinesis application
|
* Name of the Kinesis application
|
||||||
* @param recordProcessorFactory
|
* @param recordProcessorFactory
|
||||||
* Used to get record processor instances for processing data from shards
|
* Used to get record processor instances for processing data from shards
|
||||||
|
* @param config
|
||||||
|
* Kinesis Library Configuration
|
||||||
* @param streamConfig
|
* @param streamConfig
|
||||||
* Stream configuration
|
* Stream configuration
|
||||||
* @param initialPositionInStream
|
* @param initialPositionInStream
|
||||||
|
|
@ -382,7 +387,7 @@ public class Worker implements Runnable {
|
||||||
*/
|
*/
|
||||||
// NOTE: This has package level access solely for testing
|
// NOTE: This has package level access solely for testing
|
||||||
// CHECKSTYLE:IGNORE ParameterNumber FOR NEXT 10 LINES
|
// CHECKSTYLE:IGNORE ParameterNumber FOR NEXT 10 LINES
|
||||||
Worker(String applicationName, IRecordProcessorFactory recordProcessorFactory, StreamConfig streamConfig,
|
Worker(String applicationName, IRecordProcessorFactory recordProcessorFactory, KinesisClientLibConfiguration config, StreamConfig streamConfig,
|
||||||
InitialPositionInStreamExtended initialPositionInStream, long parentShardPollIntervalMillis,
|
InitialPositionInStreamExtended initialPositionInStream, long parentShardPollIntervalMillis,
|
||||||
long shardSyncIdleTimeMillis, boolean cleanupLeasesUponShardCompletion, ICheckpoint checkpoint,
|
long shardSyncIdleTimeMillis, boolean cleanupLeasesUponShardCompletion, ICheckpoint checkpoint,
|
||||||
KinesisClientLibLeaseCoordinator leaseCoordinator, ExecutorService execService,
|
KinesisClientLibLeaseCoordinator leaseCoordinator, ExecutorService execService,
|
||||||
|
|
@ -391,6 +396,7 @@ public class Worker implements Runnable {
|
||||||
Optional<Integer> retryGetRecordsInSeconds, Optional<Integer> maxGetRecordsThreadPool) {
|
Optional<Integer> retryGetRecordsInSeconds, Optional<Integer> maxGetRecordsThreadPool) {
|
||||||
this.applicationName = applicationName;
|
this.applicationName = applicationName;
|
||||||
this.recordProcessorFactory = recordProcessorFactory;
|
this.recordProcessorFactory = recordProcessorFactory;
|
||||||
|
this.config = config;
|
||||||
this.streamConfig = streamConfig;
|
this.streamConfig = streamConfig;
|
||||||
this.initialPosition = initialPositionInStream;
|
this.initialPosition = initialPositionInStream;
|
||||||
this.parentShardPollIntervalMillis = parentShardPollIntervalMillis;
|
this.parentShardPollIntervalMillis = parentShardPollIntervalMillis;
|
||||||
|
|
@ -411,7 +417,6 @@ public class Worker implements Runnable {
|
||||||
this.maxGetRecordsThreadPool = maxGetRecordsThreadPool;
|
this.maxGetRecordsThreadPool = maxGetRecordsThreadPool;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* @return the applicationName
|
* @return the applicationName
|
||||||
*/
|
*/
|
||||||
|
|
@ -819,11 +824,11 @@ public class Worker implements Runnable {
|
||||||
*
|
*
|
||||||
* @param shardInfo
|
* @param shardInfo
|
||||||
* Kinesis shard info
|
* Kinesis shard info
|
||||||
* @param factory
|
* @param processorFactory
|
||||||
* RecordProcessor factory
|
* RecordProcessor factory
|
||||||
* @return ShardConsumer for the shard
|
* @return ShardConsumer for the shard
|
||||||
*/
|
*/
|
||||||
ShardConsumer createOrGetShardConsumer(ShardInfo shardInfo, IRecordProcessorFactory factory) {
|
ShardConsumer createOrGetShardConsumer(ShardInfo shardInfo, IRecordProcessorFactory processorFactory) {
|
||||||
ShardConsumer consumer = shardInfoShardConsumerMap.get(shardInfo);
|
ShardConsumer consumer = shardInfoShardConsumerMap.get(shardInfo);
|
||||||
// Instantiate a new consumer if we don't have one, or the one we
|
// Instantiate a new consumer if we don't have one, or the one we
|
||||||
// had was from an earlier
|
// had was from an earlier
|
||||||
|
|
@ -832,20 +837,30 @@ public class Worker implements Runnable {
|
||||||
// completely processed (shutdown reason terminate).
|
// completely processed (shutdown reason terminate).
|
||||||
if ((consumer == null)
|
if ((consumer == null)
|
||||||
|| (consumer.isShutdown() && consumer.getShutdownReason().equals(ShutdownReason.ZOMBIE))) {
|
|| (consumer.isShutdown() && consumer.getShutdownReason().equals(ShutdownReason.ZOMBIE))) {
|
||||||
consumer = buildConsumer(shardInfo, factory);
|
consumer = buildConsumer(shardInfo, processorFactory);
|
||||||
shardInfoShardConsumerMap.put(shardInfo, consumer);
|
shardInfoShardConsumerMap.put(shardInfo, consumer);
|
||||||
wlog.infoForce("Created new shardConsumer for : " + shardInfo);
|
wlog.infoForce("Created new shardConsumer for : " + shardInfo);
|
||||||
}
|
}
|
||||||
return consumer;
|
return consumer;
|
||||||
}
|
}
|
||||||
|
|
||||||
protected ShardConsumer buildConsumer(ShardInfo shardInfo, IRecordProcessorFactory factory) {
|
protected ShardConsumer buildConsumer(ShardInfo shardInfo, IRecordProcessorFactory processorFactory) {
|
||||||
IRecordProcessor recordProcessor = factory.createProcessor();
|
IRecordProcessor recordProcessor = processorFactory.createProcessor();
|
||||||
|
|
||||||
return new ShardConsumer(shardInfo, streamConfig, checkpointTracker, recordProcessor,
|
return new ShardConsumer(shardInfo,
|
||||||
leaseCoordinator.getLeaseManager(), parentShardPollIntervalMillis, cleanupLeasesUponShardCompletion,
|
streamConfig,
|
||||||
executorService, metricsFactory, taskBackoffTimeMillis,
|
checkpointTracker,
|
||||||
skipShardSyncAtWorkerInitializationIfLeasesExist, retryGetRecordsInSeconds, maxGetRecordsThreadPool);
|
recordProcessor,
|
||||||
|
leaseCoordinator.getLeaseManager(),
|
||||||
|
parentShardPollIntervalMillis,
|
||||||
|
cleanupLeasesUponShardCompletion,
|
||||||
|
executorService,
|
||||||
|
metricsFactory,
|
||||||
|
taskBackoffTimeMillis,
|
||||||
|
skipShardSyncAtWorkerInitializationIfLeasesExist,
|
||||||
|
retryGetRecordsInSeconds,
|
||||||
|
maxGetRecordsThreadPool,
|
||||||
|
config);
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -1049,6 +1064,7 @@ public class Worker implements Runnable {
|
||||||
public static class Builder {
|
public static class Builder {
|
||||||
|
|
||||||
private IRecordProcessorFactory recordProcessorFactory;
|
private IRecordProcessorFactory recordProcessorFactory;
|
||||||
|
private RecordsFetcherFactory recordsFetcherFactory;
|
||||||
private KinesisClientLibConfiguration config;
|
private KinesisClientLibConfiguration config;
|
||||||
private AmazonKinesis kinesisClient;
|
private AmazonKinesis kinesisClient;
|
||||||
private AmazonDynamoDB dynamoDBClient;
|
private AmazonDynamoDB dynamoDBClient;
|
||||||
|
|
@ -1244,6 +1260,7 @@ public class Worker implements Runnable {
|
||||||
|
|
||||||
return new Worker(config.getApplicationName(),
|
return new Worker(config.getApplicationName(),
|
||||||
recordProcessorFactory,
|
recordProcessorFactory,
|
||||||
|
config,
|
||||||
new StreamConfig(new KinesisProxyFactory(config.getKinesisCredentialsProvider(),
|
new StreamConfig(new KinesisProxyFactory(config.getKinesisCredentialsProvider(),
|
||||||
kinesisClient).getProxy(config.getStreamName()),
|
kinesisClient).getProxy(config.getStreamName()),
|
||||||
config.getMaxRecords(),
|
config.getMaxRecords(),
|
||||||
|
|
|
||||||
|
|
@ -14,17 +14,25 @@
|
||||||
*/
|
*/
|
||||||
package com.amazonaws.services.kinesis.clientlibrary.types;
|
package com.amazonaws.services.kinesis.clientlibrary.types;
|
||||||
|
|
||||||
|
import java.time.Duration;
|
||||||
|
import java.time.Instant;
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
|
|
||||||
import com.amazonaws.services.kinesis.clientlibrary.interfaces.IRecordProcessorCheckpointer;
|
import com.amazonaws.services.kinesis.clientlibrary.interfaces.IRecordProcessorCheckpointer;
|
||||||
import com.amazonaws.services.kinesis.model.Record;
|
import com.amazonaws.services.kinesis.model.Record;
|
||||||
|
|
||||||
|
import lombok.Getter;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Container for the parameters to the IRecordProcessor's
|
* Container for the parameters to the IRecordProcessor's
|
||||||
* {@link com.amazonaws.services.kinesis.clientlibrary.interfaces.v2.IRecordProcessor#processRecords(
|
* {@link com.amazonaws.services.kinesis.clientlibrary.interfaces.v2.IRecordProcessor#processRecords(
|
||||||
* ProcessRecordsInput processRecordsInput) processRecords} method.
|
* ProcessRecordsInput processRecordsInput) processRecords} method.
|
||||||
*/
|
*/
|
||||||
public class ProcessRecordsInput {
|
public class ProcessRecordsInput {
|
||||||
|
@Getter
|
||||||
|
private Instant cacheEntryTime;
|
||||||
|
@Getter
|
||||||
|
private Instant cacheExitTime;
|
||||||
private List<Record> records;
|
private List<Record> records;
|
||||||
private IRecordProcessorCheckpointer checkpointer;
|
private IRecordProcessorCheckpointer checkpointer;
|
||||||
private Long millisBehindLatest;
|
private Long millisBehindLatest;
|
||||||
|
|
@ -96,4 +104,21 @@ public class ProcessRecordsInput {
|
||||||
this.millisBehindLatest = millisBehindLatest;
|
this.millisBehindLatest = millisBehindLatest;
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public ProcessRecordsInput withCacheEntryTime(Instant cacheEntryTime) {
|
||||||
|
this.cacheEntryTime = cacheEntryTime;
|
||||||
|
return this;
|
||||||
|
}
|
||||||
|
|
||||||
|
public ProcessRecordsInput withCacheExitTime(Instant cacheExitTime) {
|
||||||
|
this.cacheExitTime = cacheExitTime;
|
||||||
|
return this;
|
||||||
|
}
|
||||||
|
|
||||||
|
public Duration getTimeSpentInCache() {
|
||||||
|
if (cacheEntryTime == null || cacheExitTime == null) {
|
||||||
|
return Duration.ZERO;
|
||||||
|
}
|
||||||
|
return Duration.between(cacheEntryTime, cacheExitTime);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -14,29 +14,6 @@
|
||||||
*/
|
*/
|
||||||
package com.amazonaws.services.kinesis.clientlibrary.lib.worker;
|
package com.amazonaws.services.kinesis.clientlibrary.lib.worker;
|
||||||
|
|
||||||
import com.amazonaws.services.kinesis.clientlibrary.proxies.IKinesisProxy;
|
|
||||||
import com.amazonaws.services.kinesis.model.GetRecordsResult;
|
|
||||||
import com.google.common.util.concurrent.ThreadFactoryBuilder;
|
|
||||||
import org.junit.After;
|
|
||||||
import org.junit.Before;
|
|
||||||
import org.junit.Ignore;
|
|
||||||
import org.junit.Test;
|
|
||||||
import org.junit.runner.RunWith;
|
|
||||||
import org.mockito.Mock;
|
|
||||||
import org.mockito.runners.MockitoJUnitRunner;
|
|
||||||
|
|
||||||
import java.util.concurrent.CompletionService;
|
|
||||||
import java.util.concurrent.ExecutionException;
|
|
||||||
import java.util.concurrent.ExecutorCompletionService;
|
|
||||||
import java.util.concurrent.ExecutorService;
|
|
||||||
import java.util.concurrent.Future;
|
|
||||||
import java.util.concurrent.LinkedBlockingQueue;
|
|
||||||
import java.util.concurrent.RejectedExecutionHandler;
|
|
||||||
import java.util.concurrent.ThreadPoolExecutor;
|
|
||||||
import java.util.concurrent.TimeUnit;
|
|
||||||
import java.util.function.Supplier;
|
|
||||||
|
|
||||||
|
|
||||||
import static org.hamcrest.CoreMatchers.nullValue;
|
import static org.hamcrest.CoreMatchers.nullValue;
|
||||||
import static org.hamcrest.MatcherAssert.assertThat;
|
import static org.hamcrest.MatcherAssert.assertThat;
|
||||||
import static org.hamcrest.core.IsEqual.equalTo;
|
import static org.hamcrest.core.IsEqual.equalTo;
|
||||||
|
|
@ -50,6 +27,29 @@ import static org.mockito.Mockito.spy;
|
||||||
import static org.mockito.Mockito.verify;
|
import static org.mockito.Mockito.verify;
|
||||||
import static org.mockito.Mockito.when;
|
import static org.mockito.Mockito.when;
|
||||||
|
|
||||||
|
import java.util.concurrent.CompletionService;
|
||||||
|
import java.util.concurrent.ExecutionException;
|
||||||
|
import java.util.concurrent.ExecutorCompletionService;
|
||||||
|
import java.util.concurrent.ExecutorService;
|
||||||
|
import java.util.concurrent.Future;
|
||||||
|
import java.util.concurrent.LinkedBlockingQueue;
|
||||||
|
import java.util.concurrent.RejectedExecutionHandler;
|
||||||
|
import java.util.concurrent.ThreadPoolExecutor;
|
||||||
|
import java.util.concurrent.TimeUnit;
|
||||||
|
import java.util.function.Supplier;
|
||||||
|
|
||||||
|
import org.junit.After;
|
||||||
|
import org.junit.Before;
|
||||||
|
import org.junit.Ignore;
|
||||||
|
import org.junit.Test;
|
||||||
|
import org.junit.runner.RunWith;
|
||||||
|
import org.mockito.Mock;
|
||||||
|
import org.mockito.runners.MockitoJUnitRunner;
|
||||||
|
|
||||||
|
import com.amazonaws.services.kinesis.clientlibrary.proxies.IKinesisProxy;
|
||||||
|
import com.amazonaws.services.kinesis.model.GetRecordsResult;
|
||||||
|
import com.google.common.util.concurrent.ThreadFactoryBuilder;
|
||||||
|
|
||||||
@RunWith(MockitoJUnitRunner.class)
|
@RunWith(MockitoJUnitRunner.class)
|
||||||
public class AsynchronousGetRecordsRetrievalStrategyIntegrationTest {
|
public class AsynchronousGetRecordsRetrievalStrategyIntegrationTest {
|
||||||
|
|
||||||
|
|
@ -125,7 +125,6 @@ public class AsynchronousGetRecordsRetrievalStrategyIntegrationTest {
|
||||||
@Test
|
@Test
|
||||||
@Ignore
|
@Ignore
|
||||||
public void testInterrupted() throws InterruptedException, ExecutionException {
|
public void testInterrupted() throws InterruptedException, ExecutionException {
|
||||||
|
|
||||||
Future<DataFetcherResult> mockFuture = mock(Future.class);
|
Future<DataFetcherResult> mockFuture = mock(Future.class);
|
||||||
when(completionService.submit(any())).thenReturn(mockFuture);
|
when(completionService.submit(any())).thenReturn(mockFuture);
|
||||||
when(completionService.poll()).thenReturn(mockFuture);
|
when(completionService.poll()).thenReturn(mockFuture);
|
||||||
|
|
|
||||||
|
|
@ -0,0 +1,84 @@
|
||||||
|
/*
|
||||||
|
* Copyright 2017 Amazon.com, Inc. or its affiliates. All Rights Reserved.
|
||||||
|
*
|
||||||
|
* Licensed under the Amazon Software License (the "License").
|
||||||
|
* You may not use this file except in compliance with the License.
|
||||||
|
* A copy of the License is located at
|
||||||
|
*
|
||||||
|
* http://aws.amazon.com/asl/
|
||||||
|
*
|
||||||
|
* or in the "license" file accompanying this file. This file is distributed
|
||||||
|
* on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
|
||||||
|
* express or implied. See the License for the specific language governing
|
||||||
|
* permissions and limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package com.amazonaws.services.kinesis.clientlibrary.lib.worker;
|
||||||
|
|
||||||
|
import static org.junit.Assert.assertEquals;
|
||||||
|
import static org.junit.Assert.assertNull;
|
||||||
|
import static org.mockito.Matchers.eq;
|
||||||
|
import static org.mockito.Mockito.when;
|
||||||
|
|
||||||
|
import java.time.Duration;
|
||||||
|
import java.util.ArrayList;
|
||||||
|
import java.util.List;
|
||||||
|
|
||||||
|
import org.junit.Before;
|
||||||
|
import org.junit.Test;
|
||||||
|
import org.junit.runner.RunWith;
|
||||||
|
import org.mockito.Mock;
|
||||||
|
import org.mockito.runners.MockitoJUnitRunner;
|
||||||
|
|
||||||
|
import com.amazonaws.services.kinesis.clientlibrary.types.ProcessRecordsInput;
|
||||||
|
import com.amazonaws.services.kinesis.model.GetRecordsResult;
|
||||||
|
import com.amazonaws.services.kinesis.model.Record;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Test class for the BlockingGetRecordsCache class.
|
||||||
|
*/
|
||||||
|
@RunWith(MockitoJUnitRunner.class)
|
||||||
|
public class BlockingGetRecordsCacheTest {
|
||||||
|
private static final int MAX_RECORDS_PER_COUNT = 10_000;
|
||||||
|
private static final long IDLE_MILLIS_BETWEEN_CALLS = 500L;
|
||||||
|
|
||||||
|
@Mock
|
||||||
|
private GetRecordsRetrievalStrategy getRecordsRetrievalStrategy;
|
||||||
|
@Mock
|
||||||
|
private GetRecordsResult getRecordsResult;
|
||||||
|
|
||||||
|
private List<Record> records;
|
||||||
|
private BlockingGetRecordsCache blockingGetRecordsCache;
|
||||||
|
|
||||||
|
@Before
|
||||||
|
public void setup() {
|
||||||
|
records = new ArrayList<>();
|
||||||
|
blockingGetRecordsCache = new BlockingGetRecordsCache(MAX_RECORDS_PER_COUNT, getRecordsRetrievalStrategy, IDLE_MILLIS_BETWEEN_CALLS);
|
||||||
|
|
||||||
|
when(getRecordsRetrievalStrategy.getRecords(eq(MAX_RECORDS_PER_COUNT))).thenReturn(getRecordsResult);
|
||||||
|
when(getRecordsResult.getRecords()).thenReturn(records);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testGetNextRecordsWithNoRecords() {
|
||||||
|
ProcessRecordsInput result = blockingGetRecordsCache.getNextResult();
|
||||||
|
|
||||||
|
assertEquals(result.getRecords(), records);
|
||||||
|
assertNull(result.getCacheEntryTime());
|
||||||
|
assertNull(result.getCacheExitTime());
|
||||||
|
assertEquals(result.getTimeSpentInCache(), Duration.ZERO);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testGetNextRecordsWithRecords() {
|
||||||
|
Record record = new Record();
|
||||||
|
records.add(record);
|
||||||
|
records.add(record);
|
||||||
|
records.add(record);
|
||||||
|
records.add(record);
|
||||||
|
|
||||||
|
ProcessRecordsInput result = blockingGetRecordsCache.getNextResult();
|
||||||
|
|
||||||
|
assertEquals(result.getRecords(), records);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
@ -55,6 +55,8 @@ public class ConsumerStatesTest {
|
||||||
@Mock
|
@Mock
|
||||||
private IRecordProcessor recordProcessor;
|
private IRecordProcessor recordProcessor;
|
||||||
@Mock
|
@Mock
|
||||||
|
private KinesisClientLibConfiguration config;
|
||||||
|
@Mock
|
||||||
private RecordProcessorCheckpointer recordProcessorCheckpointer;
|
private RecordProcessorCheckpointer recordProcessorCheckpointer;
|
||||||
@Mock
|
@Mock
|
||||||
private ExecutorService executorService;
|
private ExecutorService executorService;
|
||||||
|
|
@ -75,7 +77,7 @@ public class ConsumerStatesTest {
|
||||||
@Mock
|
@Mock
|
||||||
private InitialPositionInStreamExtended initialPositionInStream;
|
private InitialPositionInStreamExtended initialPositionInStream;
|
||||||
@Mock
|
@Mock
|
||||||
private GetRecordsRetrievalStrategy getRecordsRetrievalStrategy;
|
private GetRecordsCache getRecordsCache;
|
||||||
|
|
||||||
private long parentShardPollIntervalMillis = 0xCAFE;
|
private long parentShardPollIntervalMillis = 0xCAFE;
|
||||||
private boolean cleanupLeasesOfCompletedShards = true;
|
private boolean cleanupLeasesOfCompletedShards = true;
|
||||||
|
|
@ -98,7 +100,7 @@ public class ConsumerStatesTest {
|
||||||
when(consumer.isCleanupLeasesOfCompletedShards()).thenReturn(cleanupLeasesOfCompletedShards);
|
when(consumer.isCleanupLeasesOfCompletedShards()).thenReturn(cleanupLeasesOfCompletedShards);
|
||||||
when(consumer.getTaskBackoffTimeMillis()).thenReturn(taskBackoffTimeMillis);
|
when(consumer.getTaskBackoffTimeMillis()).thenReturn(taskBackoffTimeMillis);
|
||||||
when(consumer.getShutdownReason()).thenReturn(reason);
|
when(consumer.getShutdownReason()).thenReturn(reason);
|
||||||
when(consumer.getGetRecordsRetrievalStrategy()).thenReturn(getRecordsRetrievalStrategy);
|
when(consumer.getGetRecordsCache()).thenReturn(getRecordsCache);
|
||||||
}
|
}
|
||||||
|
|
||||||
private static final Class<ILeaseManager<KinesisClientLease>> LEASE_MANAGER_CLASS = (Class<ILeaseManager<KinesisClientLease>>) (Class<?>) ILeaseManager.class;
|
private static final Class<ILeaseManager<KinesisClientLease>> LEASE_MANAGER_CLASS = (Class<ILeaseManager<KinesisClientLease>>) (Class<?>) ILeaseManager.class;
|
||||||
|
|
@ -207,6 +209,33 @@ public class ConsumerStatesTest {
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void processingStateRecordsFetcher() {
|
||||||
|
|
||||||
|
ConsumerState state = ShardConsumerState.PROCESSING.getConsumerState();
|
||||||
|
ITask task = state.createTask(consumer);
|
||||||
|
|
||||||
|
assertThat(task, procTask(ShardInfo.class, "shardInfo", equalTo(shardInfo)));
|
||||||
|
assertThat(task, procTask(IRecordProcessor.class, "recordProcessor", equalTo(recordProcessor)));
|
||||||
|
assertThat(task, procTask(RecordProcessorCheckpointer.class, "recordProcessorCheckpointer",
|
||||||
|
equalTo(recordProcessorCheckpointer)));
|
||||||
|
assertThat(task, procTask(KinesisDataFetcher.class, "dataFetcher", equalTo(dataFetcher)));
|
||||||
|
assertThat(task, procTask(StreamConfig.class, "streamConfig", equalTo(streamConfig)));
|
||||||
|
assertThat(task, procTask(Long.class, "backoffTimeMillis", equalTo(taskBackoffTimeMillis)));
|
||||||
|
|
||||||
|
assertThat(state.successTransition(), equalTo(ShardConsumerState.PROCESSING.getConsumerState()));
|
||||||
|
|
||||||
|
assertThat(state.shutdownTransition(ShutdownReason.ZOMBIE),
|
||||||
|
equalTo(ShardConsumerState.SHUTTING_DOWN.getConsumerState()));
|
||||||
|
assertThat(state.shutdownTransition(ShutdownReason.TERMINATE),
|
||||||
|
equalTo(ShardConsumerState.SHUTTING_DOWN.getConsumerState()));
|
||||||
|
assertThat(state.shutdownTransition(ShutdownReason.REQUESTED),
|
||||||
|
equalTo(ShardConsumerState.SHUTDOWN_REQUESTED.getConsumerState()));
|
||||||
|
|
||||||
|
assertThat(state.getState(), equalTo(ShardConsumerState.PROCESSING));
|
||||||
|
assertThat(state.getTaskType(), equalTo(TaskType.PROCESS));
|
||||||
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
public void shutdownRequestState() {
|
public void shutdownRequestState() {
|
||||||
ConsumerState state = ShardConsumerState.SHUTDOWN_REQUESTED.getConsumerState();
|
ConsumerState state = ShardConsumerState.SHUTDOWN_REQUESTED.getConsumerState();
|
||||||
|
|
@ -313,7 +342,7 @@ public class ConsumerStatesTest {
|
||||||
}
|
}
|
||||||
|
|
||||||
static <ValueType> ReflectionPropertyMatcher<ShutdownTask, ValueType> shutdownTask(Class<ValueType> valueTypeClass,
|
static <ValueType> ReflectionPropertyMatcher<ShutdownTask, ValueType> shutdownTask(Class<ValueType> valueTypeClass,
|
||||||
String propertyName, Matcher<ValueType> matcher) {
|
String propertyName, Matcher<ValueType> matcher) {
|
||||||
return taskWith(ShutdownTask.class, valueTypeClass, propertyName, matcher);
|
return taskWith(ShutdownTask.class, valueTypeClass, propertyName, matcher);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -323,17 +352,17 @@ public class ConsumerStatesTest {
|
||||||
}
|
}
|
||||||
|
|
||||||
static <ValueType> ReflectionPropertyMatcher<ProcessTask, ValueType> procTask(Class<ValueType> valueTypeClass,
|
static <ValueType> ReflectionPropertyMatcher<ProcessTask, ValueType> procTask(Class<ValueType> valueTypeClass,
|
||||||
String propertyName, Matcher<ValueType> matcher) {
|
String propertyName, Matcher<ValueType> matcher) {
|
||||||
return taskWith(ProcessTask.class, valueTypeClass, propertyName, matcher);
|
return taskWith(ProcessTask.class, valueTypeClass, propertyName, matcher);
|
||||||
}
|
}
|
||||||
|
|
||||||
static <ValueType> ReflectionPropertyMatcher<InitializeTask, ValueType> initTask(Class<ValueType> valueTypeClass,
|
static <ValueType> ReflectionPropertyMatcher<InitializeTask, ValueType> initTask(Class<ValueType> valueTypeClass,
|
||||||
String propertyName, Matcher<ValueType> matcher) {
|
String propertyName, Matcher<ValueType> matcher) {
|
||||||
return taskWith(InitializeTask.class, valueTypeClass, propertyName, matcher);
|
return taskWith(InitializeTask.class, valueTypeClass, propertyName, matcher);
|
||||||
}
|
}
|
||||||
|
|
||||||
static <TaskType, ValueType> ReflectionPropertyMatcher<TaskType, ValueType> taskWith(Class<TaskType> taskTypeClass,
|
static <TaskType, ValueType> ReflectionPropertyMatcher<TaskType, ValueType> taskWith(Class<TaskType> taskTypeClass,
|
||||||
Class<ValueType> valueTypeClass, String propertyName, Matcher<ValueType> matcher) {
|
Class<ValueType> valueTypeClass, String propertyName, Matcher<ValueType> matcher) {
|
||||||
return new ReflectionPropertyMatcher<>(taskTypeClass, valueTypeClass, matcher, propertyName);
|
return new ReflectionPropertyMatcher<>(taskTypeClass, valueTypeClass, matcher, propertyName);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -346,7 +375,7 @@ public class ConsumerStatesTest {
|
||||||
private final Field matchingField;
|
private final Field matchingField;
|
||||||
|
|
||||||
private ReflectionPropertyMatcher(Class<TaskType> taskTypeClass, Class<ValueType> valueTypeClass,
|
private ReflectionPropertyMatcher(Class<TaskType> taskTypeClass, Class<ValueType> valueTypeClass,
|
||||||
Matcher<ValueType> matcher, String propertyName) {
|
Matcher<ValueType> matcher, String propertyName) {
|
||||||
this.taskTypeClass = taskTypeClass;
|
this.taskTypeClass = taskTypeClass;
|
||||||
this.valueTypeClazz = valueTypeClass;
|
this.valueTypeClazz = valueTypeClass;
|
||||||
this.matcher = matcher;
|
this.matcher = matcher;
|
||||||
|
|
|
||||||
|
|
@ -19,7 +19,7 @@ import static org.junit.Assert.assertNull;
|
||||||
import static org.junit.Assert.assertTrue;
|
import static org.junit.Assert.assertTrue;
|
||||||
import static org.junit.Assert.fail;
|
import static org.junit.Assert.fail;
|
||||||
|
|
||||||
import junit.framework.Assert;
|
import java.util.Date;
|
||||||
|
|
||||||
import org.junit.Test;
|
import org.junit.Test;
|
||||||
import org.mockito.Mockito;
|
import org.mockito.Mockito;
|
||||||
|
|
@ -35,7 +35,7 @@ import com.amazonaws.services.kinesis.clientlibrary.interfaces.IRecordProcessorF
|
||||||
import com.amazonaws.services.kinesis.metrics.interfaces.MetricsLevel;
|
import com.amazonaws.services.kinesis.metrics.interfaces.MetricsLevel;
|
||||||
import com.google.common.collect.ImmutableSet;
|
import com.google.common.collect.ImmutableSet;
|
||||||
|
|
||||||
import java.util.Date;
|
import junit.framework.Assert;
|
||||||
|
|
||||||
public class KinesisClientLibConfigurationTest {
|
public class KinesisClientLibConfigurationTest {
|
||||||
private static final long INVALID_LONG = 0L;
|
private static final long INVALID_LONG = 0L;
|
||||||
|
|
@ -95,7 +95,8 @@ public class KinesisClientLibConfigurationTest {
|
||||||
// Try each argument at one time.
|
// Try each argument at one time.
|
||||||
KinesisClientLibConfiguration config = null;
|
KinesisClientLibConfiguration config = null;
|
||||||
long[] longValues =
|
long[] longValues =
|
||||||
{ TEST_VALUE_LONG, TEST_VALUE_LONG, TEST_VALUE_LONG, TEST_VALUE_LONG, TEST_VALUE_LONG, TEST_VALUE_LONG, TEST_VALUE_LONG };
|
{ TEST_VALUE_LONG, TEST_VALUE_LONG, TEST_VALUE_LONG, TEST_VALUE_LONG, TEST_VALUE_LONG, TEST_VALUE_LONG,
|
||||||
|
TEST_VALUE_LONG };
|
||||||
for (int i = 0; i < PARAMETER_COUNT; i++) {
|
for (int i = 0; i < PARAMETER_COUNT; i++) {
|
||||||
longValues[i] = INVALID_LONG;
|
longValues[i] = INVALID_LONG;
|
||||||
try {
|
try {
|
||||||
|
|
@ -300,30 +301,30 @@ public class KinesisClientLibConfigurationTest {
|
||||||
Mockito.mock(AWSCredentialsProvider.class);
|
Mockito.mock(AWSCredentialsProvider.class);
|
||||||
try {
|
try {
|
||||||
new KinesisClientLibConfiguration(TEST_STRING,
|
new KinesisClientLibConfiguration(TEST_STRING,
|
||||||
TEST_STRING,
|
TEST_STRING,
|
||||||
TEST_STRING,
|
TEST_STRING,
|
||||||
TEST_STRING,
|
TEST_STRING,
|
||||||
null,
|
null,
|
||||||
null,
|
null,
|
||||||
null,
|
null,
|
||||||
null,
|
null,
|
||||||
TEST_VALUE_LONG,
|
TEST_VALUE_LONG,
|
||||||
TEST_STRING,
|
TEST_STRING,
|
||||||
3,
|
3,
|
||||||
TEST_VALUE_LONG,
|
TEST_VALUE_LONG,
|
||||||
false,
|
false,
|
||||||
TEST_VALUE_LONG,
|
TEST_VALUE_LONG,
|
||||||
TEST_VALUE_LONG,
|
TEST_VALUE_LONG,
|
||||||
true,
|
true,
|
||||||
new ClientConfiguration(),
|
new ClientConfiguration(),
|
||||||
new ClientConfiguration(),
|
new ClientConfiguration(),
|
||||||
new ClientConfiguration(),
|
new ClientConfiguration(),
|
||||||
TEST_VALUE_LONG,
|
TEST_VALUE_LONG,
|
||||||
TEST_VALUE_LONG,
|
TEST_VALUE_LONG,
|
||||||
1,
|
1,
|
||||||
skipCheckpointValidationValue,
|
skipCheckpointValidationValue,
|
||||||
"abcd",
|
"abcd",
|
||||||
TEST_VALUE_LONG);
|
TEST_VALUE_LONG);
|
||||||
Assert.fail("No expected Exception is thrown.");
|
Assert.fail("No expected Exception is thrown.");
|
||||||
} catch(IllegalArgumentException e) {
|
} catch(IllegalArgumentException e) {
|
||||||
System.out.println(e.getMessage());
|
System.out.println(e.getMessage());
|
||||||
|
|
|
||||||
|
|
@ -15,9 +15,10 @@
|
||||||
package com.amazonaws.services.kinesis.clientlibrary.lib.worker;
|
package com.amazonaws.services.kinesis.clientlibrary.lib.worker;
|
||||||
|
|
||||||
import static org.hamcrest.CoreMatchers.equalTo;
|
import static org.hamcrest.CoreMatchers.equalTo;
|
||||||
import static org.hamcrest.CoreMatchers.not;
|
import static org.hamcrest.CoreMatchers.notNullValue;
|
||||||
import static org.hamcrest.CoreMatchers.nullValue;
|
import static org.hamcrest.CoreMatchers.nullValue;
|
||||||
import static org.hamcrest.MatcherAssert.assertThat;
|
import static org.hamcrest.MatcherAssert.assertThat;
|
||||||
|
import static org.hamcrest.collection.IsEmptyCollection.empty;
|
||||||
import static org.mockito.Matchers.anyInt;
|
import static org.mockito.Matchers.anyInt;
|
||||||
import static org.mockito.Matchers.anyString;
|
import static org.mockito.Matchers.anyString;
|
||||||
import static org.mockito.Matchers.eq;
|
import static org.mockito.Matchers.eq;
|
||||||
|
|
@ -37,11 +38,10 @@ import java.util.List;
|
||||||
import org.junit.Assert;
|
import org.junit.Assert;
|
||||||
import org.junit.BeforeClass;
|
import org.junit.BeforeClass;
|
||||||
import org.junit.Test;
|
import org.junit.Test;
|
||||||
|
import org.junit.runner.RunWith;
|
||||||
|
import org.mockito.Mock;
|
||||||
|
import org.mockito.runners.MockitoJUnitRunner;
|
||||||
|
|
||||||
import com.amazonaws.services.kinesis.model.GetRecordsResult;
|
|
||||||
import com.amazonaws.services.kinesis.model.Record;
|
|
||||||
import com.amazonaws.services.kinesis.model.ResourceNotFoundException;
|
|
||||||
import com.amazonaws.services.kinesis.model.ShardIteratorType;
|
|
||||||
import com.amazonaws.services.kinesis.clientlibrary.exceptions.KinesisClientLibException;
|
import com.amazonaws.services.kinesis.clientlibrary.exceptions.KinesisClientLibException;
|
||||||
import com.amazonaws.services.kinesis.clientlibrary.interfaces.ICheckpoint;
|
import com.amazonaws.services.kinesis.clientlibrary.interfaces.ICheckpoint;
|
||||||
import com.amazonaws.services.kinesis.clientlibrary.lib.checkpoint.SentinelCheckpoint;
|
import com.amazonaws.services.kinesis.clientlibrary.lib.checkpoint.SentinelCheckpoint;
|
||||||
|
|
@ -50,9 +50,10 @@ import com.amazonaws.services.kinesis.clientlibrary.proxies.KinesisProxy;
|
||||||
import com.amazonaws.services.kinesis.clientlibrary.types.ExtendedSequenceNumber;
|
import com.amazonaws.services.kinesis.clientlibrary.types.ExtendedSequenceNumber;
|
||||||
import com.amazonaws.services.kinesis.metrics.impl.MetricsHelper;
|
import com.amazonaws.services.kinesis.metrics.impl.MetricsHelper;
|
||||||
import com.amazonaws.services.kinesis.metrics.impl.NullMetricsFactory;
|
import com.amazonaws.services.kinesis.metrics.impl.NullMetricsFactory;
|
||||||
import org.junit.runner.RunWith;
|
import com.amazonaws.services.kinesis.model.GetRecordsResult;
|
||||||
import org.mockito.Mock;
|
import com.amazonaws.services.kinesis.model.Record;
|
||||||
import org.mockito.runners.MockitoJUnitRunner;
|
import com.amazonaws.services.kinesis.model.ResourceNotFoundException;
|
||||||
|
import com.amazonaws.services.kinesis.model.ShardIteratorType;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Unit tests for KinesisDataFetcher.
|
* Unit tests for KinesisDataFetcher.
|
||||||
|
|
@ -209,6 +210,103 @@ public class KinesisDataFetcherTest {
|
||||||
Assert.assertTrue("Shard should reach the end", dataFetcher.isShardEndReached());
|
Assert.assertTrue("Shard should reach the end", dataFetcher.isShardEndReached());
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testNonNullGetRecords() {
|
||||||
|
String nextIterator = "TestIterator";
|
||||||
|
int maxRecords = 100;
|
||||||
|
|
||||||
|
KinesisProxy mockProxy = mock(KinesisProxy.class);
|
||||||
|
doThrow(new ResourceNotFoundException("Test Exception")).when(mockProxy).get(nextIterator, maxRecords);
|
||||||
|
|
||||||
|
KinesisDataFetcher dataFetcher = new KinesisDataFetcher(mockProxy, SHARD_INFO);
|
||||||
|
dataFetcher.initialize(SentinelCheckpoint.LATEST.toString(), INITIAL_POSITION_LATEST);
|
||||||
|
|
||||||
|
DataFetcherResult dataFetcherResult = dataFetcher.getRecords(maxRecords);
|
||||||
|
|
||||||
|
assertThat(dataFetcherResult, notNullValue());
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testFetcherDoesNotAdvanceWithoutAccept() {
|
||||||
|
final String INITIAL_ITERATOR = "InitialIterator";
|
||||||
|
final String NEXT_ITERATOR_ONE = "NextIteratorOne";
|
||||||
|
final String NEXT_ITERATOR_TWO = "NextIteratorTwo";
|
||||||
|
when(kinesisProxy.getIterator(anyString(), anyString())).thenReturn(INITIAL_ITERATOR);
|
||||||
|
GetRecordsResult iteratorOneResults = mock(GetRecordsResult.class);
|
||||||
|
when(iteratorOneResults.getNextShardIterator()).thenReturn(NEXT_ITERATOR_ONE);
|
||||||
|
when(kinesisProxy.get(eq(INITIAL_ITERATOR), anyInt())).thenReturn(iteratorOneResults);
|
||||||
|
|
||||||
|
GetRecordsResult iteratorTwoResults = mock(GetRecordsResult.class);
|
||||||
|
when(kinesisProxy.get(eq(NEXT_ITERATOR_ONE), anyInt())).thenReturn(iteratorTwoResults);
|
||||||
|
when(iteratorTwoResults.getNextShardIterator()).thenReturn(NEXT_ITERATOR_TWO);
|
||||||
|
|
||||||
|
GetRecordsResult finalResult = mock(GetRecordsResult.class);
|
||||||
|
when(kinesisProxy.get(eq(NEXT_ITERATOR_TWO), anyInt())).thenReturn(finalResult);
|
||||||
|
when(finalResult.getNextShardIterator()).thenReturn(null);
|
||||||
|
|
||||||
|
KinesisDataFetcher dataFetcher = new KinesisDataFetcher(kinesisProxy, SHARD_INFO);
|
||||||
|
dataFetcher.initialize("TRIM_HORIZON",
|
||||||
|
InitialPositionInStreamExtended.newInitialPosition(InitialPositionInStream.TRIM_HORIZON));
|
||||||
|
|
||||||
|
assertNoAdvance(dataFetcher, iteratorOneResults, INITIAL_ITERATOR);
|
||||||
|
assertAdvanced(dataFetcher, iteratorOneResults, INITIAL_ITERATOR, NEXT_ITERATOR_ONE);
|
||||||
|
|
||||||
|
assertNoAdvance(dataFetcher, iteratorTwoResults, NEXT_ITERATOR_ONE);
|
||||||
|
assertAdvanced(dataFetcher, iteratorTwoResults, NEXT_ITERATOR_ONE, NEXT_ITERATOR_TWO);
|
||||||
|
|
||||||
|
assertNoAdvance(dataFetcher, finalResult, NEXT_ITERATOR_TWO);
|
||||||
|
assertAdvanced(dataFetcher, finalResult, NEXT_ITERATOR_TWO, null);
|
||||||
|
|
||||||
|
verify(kinesisProxy, times(2)).get(eq(INITIAL_ITERATOR), anyInt());
|
||||||
|
verify(kinesisProxy, times(2)).get(eq(NEXT_ITERATOR_ONE), anyInt());
|
||||||
|
verify(kinesisProxy, times(2)).get(eq(NEXT_ITERATOR_TWO), anyInt());
|
||||||
|
|
||||||
|
reset(kinesisProxy);
|
||||||
|
|
||||||
|
DataFetcherResult terminal = dataFetcher.getRecords(100);
|
||||||
|
assertThat(terminal.isShardEnd(), equalTo(true));
|
||||||
|
assertThat(terminal.getResult(), notNullValue());
|
||||||
|
GetRecordsResult terminalResult = terminal.getResult();
|
||||||
|
assertThat(terminalResult.getRecords(), notNullValue());
|
||||||
|
assertThat(terminalResult.getRecords(), empty());
|
||||||
|
assertThat(terminalResult.getNextShardIterator(), nullValue());
|
||||||
|
assertThat(terminal, equalTo(dataFetcher.TERMINAL_RESULT));
|
||||||
|
|
||||||
|
verify(kinesisProxy, never()).get(anyString(), anyInt());
|
||||||
|
}
|
||||||
|
|
||||||
|
private DataFetcherResult assertAdvanced(KinesisDataFetcher dataFetcher, GetRecordsResult expectedResult,
|
||||||
|
String previousValue, String nextValue) {
|
||||||
|
DataFetcherResult acceptResult = dataFetcher.getRecords(100);
|
||||||
|
assertThat(acceptResult.getResult(), equalTo(expectedResult));
|
||||||
|
|
||||||
|
assertThat(dataFetcher.getNextIterator(), equalTo(previousValue));
|
||||||
|
assertThat(dataFetcher.isShardEndReached(), equalTo(false));
|
||||||
|
|
||||||
|
assertThat(acceptResult.accept(), equalTo(expectedResult));
|
||||||
|
assertThat(dataFetcher.getNextIterator(), equalTo(nextValue));
|
||||||
|
if (nextValue == null) {
|
||||||
|
assertThat(dataFetcher.isShardEndReached(), equalTo(true));
|
||||||
|
}
|
||||||
|
|
||||||
|
verify(kinesisProxy, times(2)).get(eq(previousValue), anyInt());
|
||||||
|
|
||||||
|
return acceptResult;
|
||||||
|
}
|
||||||
|
|
||||||
|
private DataFetcherResult assertNoAdvance(KinesisDataFetcher dataFetcher, GetRecordsResult expectedResult,
|
||||||
|
String previousValue) {
|
||||||
|
assertThat(dataFetcher.getNextIterator(), equalTo(previousValue));
|
||||||
|
DataFetcherResult noAcceptResult = dataFetcher.getRecords(100);
|
||||||
|
assertThat(noAcceptResult.getResult(), equalTo(expectedResult));
|
||||||
|
|
||||||
|
assertThat(dataFetcher.getNextIterator(), equalTo(previousValue));
|
||||||
|
|
||||||
|
verify(kinesisProxy).get(eq(previousValue), anyInt());
|
||||||
|
|
||||||
|
return noAcceptResult;
|
||||||
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
public void testFetcherDoesNotAdvanceWithoutAccept() {
|
public void testFetcherDoesNotAdvanceWithoutAccept() {
|
||||||
final String INITIAL_ITERATOR = "InitialIterator";
|
final String INITIAL_ITERATOR = "InitialIterator";
|
||||||
|
|
|
||||||
|
|
@ -0,0 +1,199 @@
|
||||||
|
/*
|
||||||
|
* Copyright 2017 Amazon.com, Inc. or its affiliates. All Rights Reserved.
|
||||||
|
*
|
||||||
|
* Licensed under the Amazon Software License (the "License").
|
||||||
|
* You may not use this file except in compliance with the License.
|
||||||
|
* A copy of the License is located at
|
||||||
|
*
|
||||||
|
* http://aws.amazon.com/asl/
|
||||||
|
*
|
||||||
|
* or in the "license" file accompanying this file. This file is distributed
|
||||||
|
* on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
|
||||||
|
* express or implied. See the License for the specific language governing
|
||||||
|
* permissions and limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package com.amazonaws.services.kinesis.clientlibrary.lib.worker;
|
||||||
|
|
||||||
|
import static org.junit.Assert.assertEquals;
|
||||||
|
import static org.junit.Assert.assertFalse;
|
||||||
|
import static org.junit.Assert.assertNotEquals;
|
||||||
|
import static org.junit.Assert.assertNotNull;
|
||||||
|
import static org.junit.Assert.assertTrue;
|
||||||
|
import static org.mockito.Mockito.mock;
|
||||||
|
import static org.mockito.Mockito.spy;
|
||||||
|
import static org.mockito.Mockito.verify;
|
||||||
|
import static org.mockito.Mockito.when;
|
||||||
|
|
||||||
|
import java.nio.ByteBuffer;
|
||||||
|
import java.util.ArrayList;
|
||||||
|
import java.util.List;
|
||||||
|
import java.util.concurrent.ExecutorService;
|
||||||
|
import java.util.concurrent.Executors;
|
||||||
|
|
||||||
|
import com.amazonaws.services.kinesis.clientlibrary.types.ProcessRecordsInput;
|
||||||
|
import com.amazonaws.services.kinesis.metrics.impl.NullMetricsFactory;
|
||||||
|
import com.amazonaws.services.kinesis.model.Record;
|
||||||
|
|
||||||
|
import org.junit.After;
|
||||||
|
import org.junit.Before;
|
||||||
|
import org.junit.Test;
|
||||||
|
import org.junit.runner.RunWith;
|
||||||
|
import org.mockito.Mock;
|
||||||
|
import org.mockito.runners.MockitoJUnitRunner;
|
||||||
|
|
||||||
|
import com.amazonaws.services.kinesis.clientlibrary.proxies.IKinesisProxy;
|
||||||
|
import com.amazonaws.services.kinesis.clientlibrary.types.ProcessRecordsInput;
|
||||||
|
import com.amazonaws.services.kinesis.model.GetRecordsResult;
|
||||||
|
import com.amazonaws.services.kinesis.model.Record;
|
||||||
|
|
||||||
|
import lombok.extern.apachecommons.CommonsLog;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* These are the integration tests for the PrefetchGetRecordsCache class.
|
||||||
|
*/
|
||||||
|
@RunWith(MockitoJUnitRunner.class)
|
||||||
|
@CommonsLog
|
||||||
|
public class PrefetchGetRecordsCacheIntegrationTest {
|
||||||
|
private static final int MAX_SIZE = 3;
|
||||||
|
private static final int MAX_BYTE_SIZE = 5 * 1024 * 1024;
|
||||||
|
private static final int MAX_RECORDS_COUNT = 30_000;
|
||||||
|
private static final int MAX_RECORDS_PER_CALL = 10_000;
|
||||||
|
private static final long IDLE_MILLIS_BETWEEN_CALLS = 500L;
|
||||||
|
|
||||||
|
private PrefetchGetRecordsCache getRecordsCache;
|
||||||
|
private GetRecordsRetrievalStrategy getRecordsRetrievalStrategy;
|
||||||
|
private KinesisDataFetcher dataFetcher;
|
||||||
|
private ExecutorService executorService;
|
||||||
|
private List<Record> records;
|
||||||
|
private String operation = "ProcessTask";
|
||||||
|
|
||||||
|
@Mock
|
||||||
|
private IKinesisProxy proxy;
|
||||||
|
|
||||||
|
@Mock
|
||||||
|
private ShardInfo shardInfo;
|
||||||
|
|
||||||
|
@Before
|
||||||
|
public void setup() {
|
||||||
|
records = new ArrayList<>();
|
||||||
|
dataFetcher = new KinesisDataFetcherForTest(proxy, shardInfo);
|
||||||
|
getRecordsRetrievalStrategy = spy(new SynchronousGetRecordsRetrievalStrategy(dataFetcher));
|
||||||
|
executorService = spy(Executors.newFixedThreadPool(1));
|
||||||
|
|
||||||
|
getRecordsCache = new PrefetchGetRecordsCache(MAX_SIZE,
|
||||||
|
MAX_BYTE_SIZE,
|
||||||
|
MAX_RECORDS_COUNT,
|
||||||
|
MAX_RECORDS_PER_CALL,
|
||||||
|
getRecordsRetrievalStrategy,
|
||||||
|
executorService,
|
||||||
|
IDLE_MILLIS_BETWEEN_CALLS,
|
||||||
|
new NullMetricsFactory(),
|
||||||
|
operation);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testRollingCache() {
|
||||||
|
getRecordsCache.start();
|
||||||
|
sleep(IDLE_MILLIS_BETWEEN_CALLS);
|
||||||
|
|
||||||
|
ProcessRecordsInput processRecordsInput1 = getRecordsCache.getNextResult();
|
||||||
|
|
||||||
|
assertTrue(processRecordsInput1.getRecords().isEmpty());
|
||||||
|
assertEquals(processRecordsInput1.getMillisBehindLatest(), new Long(1000));
|
||||||
|
assertNotNull(processRecordsInput1.getCacheEntryTime());
|
||||||
|
|
||||||
|
ProcessRecordsInput processRecordsInput2 = getRecordsCache.getNextResult();
|
||||||
|
|
||||||
|
assertNotEquals(processRecordsInput1, processRecordsInput2);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testFullCache() {
|
||||||
|
getRecordsCache.start();
|
||||||
|
sleep(MAX_SIZE * IDLE_MILLIS_BETWEEN_CALLS);
|
||||||
|
|
||||||
|
assertEquals(getRecordsCache.getRecordsResultQueue.size(), MAX_SIZE);
|
||||||
|
|
||||||
|
ProcessRecordsInput processRecordsInput1 = getRecordsCache.getNextResult();
|
||||||
|
ProcessRecordsInput processRecordsInput2 = getRecordsCache.getNextResult();
|
||||||
|
|
||||||
|
assertNotEquals(processRecordsInput1, processRecordsInput2);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testDifferentShardCaches() {
|
||||||
|
ExecutorService executorService2 = spy(Executors.newFixedThreadPool(1));
|
||||||
|
KinesisDataFetcher kinesisDataFetcher = spy(new KinesisDataFetcherForTest(proxy, shardInfo));
|
||||||
|
GetRecordsRetrievalStrategy getRecordsRetrievalStrategy2 = spy(new AsynchronousGetRecordsRetrievalStrategy(kinesisDataFetcher, 5 , 5, "Test-shard"));
|
||||||
|
GetRecordsCache getRecordsCache2 = new PrefetchGetRecordsCache(
|
||||||
|
MAX_SIZE,
|
||||||
|
MAX_BYTE_SIZE,
|
||||||
|
MAX_RECORDS_COUNT,
|
||||||
|
MAX_RECORDS_PER_CALL,
|
||||||
|
getRecordsRetrievalStrategy2,
|
||||||
|
executorService2,
|
||||||
|
IDLE_MILLIS_BETWEEN_CALLS,
|
||||||
|
new NullMetricsFactory(),
|
||||||
|
operation);
|
||||||
|
|
||||||
|
getRecordsCache.start();
|
||||||
|
sleep(IDLE_MILLIS_BETWEEN_CALLS);
|
||||||
|
|
||||||
|
Record record = mock(Record.class);
|
||||||
|
ByteBuffer byteBuffer = ByteBuffer.allocate(512 * 1024);
|
||||||
|
when(record.getData()).thenReturn(byteBuffer);
|
||||||
|
|
||||||
|
records.add(record);
|
||||||
|
records.add(record);
|
||||||
|
records.add(record);
|
||||||
|
records.add(record);
|
||||||
|
getRecordsCache2.start();
|
||||||
|
|
||||||
|
sleep(IDLE_MILLIS_BETWEEN_CALLS);
|
||||||
|
|
||||||
|
ProcessRecordsInput p1 = getRecordsCache.getNextResult();
|
||||||
|
|
||||||
|
ProcessRecordsInput p2 = getRecordsCache2.getNextResult();
|
||||||
|
|
||||||
|
assertNotEquals(p1, p2);
|
||||||
|
assertTrue(p1.getRecords().isEmpty());
|
||||||
|
assertFalse(p2.getRecords().isEmpty());
|
||||||
|
assertEquals(p2.getRecords().size(), records.size());
|
||||||
|
|
||||||
|
getRecordsCache2.shutdown();
|
||||||
|
sleep(100L);
|
||||||
|
verify(executorService2).shutdownNow();
|
||||||
|
verify(getRecordsRetrievalStrategy2).shutdown();
|
||||||
|
}
|
||||||
|
|
||||||
|
@After
|
||||||
|
public void shutdown() {
|
||||||
|
getRecordsCache.shutdown();
|
||||||
|
sleep(100L);
|
||||||
|
verify(executorService).shutdownNow();
|
||||||
|
verify(getRecordsRetrievalStrategy).shutdown();
|
||||||
|
}
|
||||||
|
|
||||||
|
private void sleep(long millis) {
|
||||||
|
try {
|
||||||
|
Thread.sleep(millis);
|
||||||
|
} catch (InterruptedException e) {}
|
||||||
|
}
|
||||||
|
|
||||||
|
private class KinesisDataFetcherForTest extends KinesisDataFetcher {
|
||||||
|
public KinesisDataFetcherForTest(final IKinesisProxy kinesisProxy,
|
||||||
|
final ShardInfo shardInfo) {
|
||||||
|
super(kinesisProxy, shardInfo);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public DataFetcherResult getRecords(final int maxRecords) {
|
||||||
|
GetRecordsResult getRecordsResult = new GetRecordsResult();
|
||||||
|
getRecordsResult.setRecords(new ArrayList<>(records));
|
||||||
|
getRecordsResult.setMillisBehindLatest(1000L);
|
||||||
|
|
||||||
|
return new AdvancingResult(getRecordsResult);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
@ -0,0 +1,215 @@
|
||||||
|
/*
|
||||||
|
* Copyright 2017 Amazon.com, Inc. or its affiliates. All Rights Reserved.
|
||||||
|
*
|
||||||
|
* Licensed under the Amazon Software License (the "License").
|
||||||
|
* You may not use this file except in compliance with the License.
|
||||||
|
* A copy of the License is located at
|
||||||
|
*
|
||||||
|
* http://aws.amazon.com/asl/
|
||||||
|
*
|
||||||
|
* or in the "license" file accompanying this file. This file is distributed
|
||||||
|
* on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
|
||||||
|
* express or implied. See the License for the specific language governing
|
||||||
|
* permissions and limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package com.amazonaws.services.kinesis.clientlibrary.lib.worker;
|
||||||
|
|
||||||
|
import static org.junit.Assert.assertEquals;
|
||||||
|
import static org.junit.Assert.assertNotEquals;
|
||||||
|
import static org.junit.Assert.assertNotNull;
|
||||||
|
import static org.junit.Assert.assertTrue;
|
||||||
|
import static org.mockito.Matchers.any;
|
||||||
|
import static org.mockito.Matchers.eq;
|
||||||
|
import static org.mockito.Mockito.atLeast;
|
||||||
|
import static org.mockito.Mockito.spy;
|
||||||
|
import static org.mockito.Mockito.times;
|
||||||
|
import static org.mockito.Mockito.verify;
|
||||||
|
import static org.mockito.Mockito.when;
|
||||||
|
|
||||||
|
import java.nio.ByteBuffer;
|
||||||
|
import java.time.Duration;
|
||||||
|
import java.util.ArrayList;
|
||||||
|
import java.util.List;
|
||||||
|
import java.util.concurrent.ExecutorService;
|
||||||
|
import java.util.concurrent.Executors;
|
||||||
|
import java.util.concurrent.LinkedBlockingQueue;
|
||||||
|
import java.util.stream.IntStream;
|
||||||
|
|
||||||
|
import com.amazonaws.services.kinesis.metrics.impl.NullMetricsFactory;
|
||||||
|
import org.junit.After;
|
||||||
|
import org.junit.Before;
|
||||||
|
import org.junit.Test;
|
||||||
|
import org.junit.runner.RunWith;
|
||||||
|
import org.mockito.Mock;
|
||||||
|
import org.mockito.runners.MockitoJUnitRunner;
|
||||||
|
|
||||||
|
import com.amazonaws.services.kinesis.clientlibrary.types.ProcessRecordsInput;
|
||||||
|
import com.amazonaws.services.kinesis.model.GetRecordsResult;
|
||||||
|
import com.amazonaws.services.kinesis.model.Record;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Test class for the PrefetchGetRecordsCache class.
|
||||||
|
*/
|
||||||
|
@RunWith(MockitoJUnitRunner.class)
|
||||||
|
public class PrefetchGetRecordsCacheTest {
|
||||||
|
private static final int SIZE_512_KB = 512 * 1024;
|
||||||
|
private static final int SIZE_1_MB = 2 * SIZE_512_KB;
|
||||||
|
private static final int MAX_RECORDS_PER_CALL = 10000;
|
||||||
|
private static final int MAX_SIZE = 5;
|
||||||
|
private static final int MAX_RECORDS_COUNT = 15000;
|
||||||
|
private static final long IDLE_MILLIS_BETWEEN_CALLS = 0L;
|
||||||
|
|
||||||
|
@Mock
|
||||||
|
private GetRecordsRetrievalStrategy getRecordsRetrievalStrategy;
|
||||||
|
@Mock
|
||||||
|
private GetRecordsResult getRecordsResult;
|
||||||
|
@Mock
|
||||||
|
private Record record;
|
||||||
|
|
||||||
|
private List<Record> records;
|
||||||
|
private ExecutorService executorService;
|
||||||
|
private LinkedBlockingQueue<ProcessRecordsInput> spyQueue;
|
||||||
|
private PrefetchGetRecordsCache getRecordsCache;
|
||||||
|
private String operation = "ProcessTask";
|
||||||
|
|
||||||
|
@Before
|
||||||
|
public void setup() {
|
||||||
|
executorService = spy(Executors.newFixedThreadPool(1));
|
||||||
|
getRecordsCache = new PrefetchGetRecordsCache(
|
||||||
|
MAX_SIZE,
|
||||||
|
3 * SIZE_1_MB,
|
||||||
|
MAX_RECORDS_COUNT,
|
||||||
|
MAX_RECORDS_PER_CALL,
|
||||||
|
getRecordsRetrievalStrategy,
|
||||||
|
executorService,
|
||||||
|
IDLE_MILLIS_BETWEEN_CALLS,
|
||||||
|
new NullMetricsFactory(),
|
||||||
|
operation);
|
||||||
|
spyQueue = spy(getRecordsCache.getRecordsResultQueue);
|
||||||
|
records = spy(new ArrayList<>());
|
||||||
|
|
||||||
|
when(getRecordsRetrievalStrategy.getRecords(eq(MAX_RECORDS_PER_CALL))).thenReturn(getRecordsResult);
|
||||||
|
when(getRecordsResult.getRecords()).thenReturn(records);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testGetRecords() {
|
||||||
|
when(records.size()).thenReturn(1000);
|
||||||
|
when(record.getData()).thenReturn(createByteBufferWithSize(SIZE_512_KB));
|
||||||
|
|
||||||
|
records.add(record);
|
||||||
|
records.add(record);
|
||||||
|
records.add(record);
|
||||||
|
records.add(record);
|
||||||
|
records.add(record);
|
||||||
|
|
||||||
|
getRecordsCache.start();
|
||||||
|
ProcessRecordsInput result = getRecordsCache.getNextResult();
|
||||||
|
|
||||||
|
assertEquals(result.getRecords(), records);
|
||||||
|
|
||||||
|
verify(executorService).execute(any());
|
||||||
|
verify(getRecordsRetrievalStrategy, atLeast(1)).getRecords(eq(MAX_RECORDS_PER_CALL));
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testFullCacheByteSize() {
|
||||||
|
when(records.size()).thenReturn(500);
|
||||||
|
when(record.getData()).thenReturn(createByteBufferWithSize(SIZE_1_MB));
|
||||||
|
|
||||||
|
records.add(record);
|
||||||
|
|
||||||
|
getRecordsCache.start();
|
||||||
|
|
||||||
|
// Sleep for a few seconds for the cache to fill up.
|
||||||
|
sleep(2000);
|
||||||
|
|
||||||
|
verify(getRecordsRetrievalStrategy, times(3)).getRecords(eq(MAX_RECORDS_PER_CALL));
|
||||||
|
assertEquals(spyQueue.size(), 3);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testFullCacheRecordsCount() {
|
||||||
|
int recordsSize = 4500;
|
||||||
|
when(records.size()).thenReturn(recordsSize);
|
||||||
|
|
||||||
|
getRecordsCache.start();
|
||||||
|
|
||||||
|
sleep(2000);
|
||||||
|
|
||||||
|
int callRate = (int) Math.ceil((double) MAX_RECORDS_COUNT/recordsSize);
|
||||||
|
verify(getRecordsRetrievalStrategy, times(callRate)).getRecords(MAX_RECORDS_PER_CALL);
|
||||||
|
assertEquals(spyQueue.size(), callRate);
|
||||||
|
assertTrue(callRate < MAX_SIZE);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testFullCacheSize() {
|
||||||
|
int recordsSize = 200;
|
||||||
|
when(records.size()).thenReturn(recordsSize);
|
||||||
|
|
||||||
|
getRecordsCache.start();
|
||||||
|
|
||||||
|
// Sleep for a few seconds for the cache to fill up.
|
||||||
|
sleep(2000);
|
||||||
|
|
||||||
|
verify(getRecordsRetrievalStrategy, times(MAX_SIZE + 1)).getRecords(eq(MAX_RECORDS_PER_CALL));
|
||||||
|
assertEquals(spyQueue.size(), MAX_SIZE);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testMultipleCacheCalls() {
|
||||||
|
int recordsSize = 20;
|
||||||
|
when(record.getData()).thenReturn(createByteBufferWithSize(1024));
|
||||||
|
|
||||||
|
IntStream.range(0, recordsSize).forEach(i -> records.add(record));
|
||||||
|
|
||||||
|
getRecordsCache.start();
|
||||||
|
ProcessRecordsInput processRecordsInput = getRecordsCache.getNextResult();
|
||||||
|
|
||||||
|
verify(executorService).execute(any());
|
||||||
|
assertEquals(processRecordsInput.getRecords(), records);
|
||||||
|
assertNotNull(processRecordsInput.getCacheEntryTime());
|
||||||
|
assertNotNull(processRecordsInput.getCacheExitTime());
|
||||||
|
|
||||||
|
sleep(2000);
|
||||||
|
|
||||||
|
ProcessRecordsInput processRecordsInput2 = getRecordsCache.getNextResult();
|
||||||
|
assertNotEquals(processRecordsInput, processRecordsInput2);
|
||||||
|
assertEquals(processRecordsInput2.getRecords(), records);
|
||||||
|
assertNotEquals(processRecordsInput2.getTimeSpentInCache(), Duration.ZERO);
|
||||||
|
|
||||||
|
assertTrue(spyQueue.size() <= MAX_SIZE);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test(expected = IllegalStateException.class)
|
||||||
|
public void testGetNextRecordsWithoutStarting() {
|
||||||
|
verify(executorService, times(0)).execute(any());
|
||||||
|
getRecordsCache.getNextResult();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test(expected = IllegalStateException.class)
|
||||||
|
public void testCallAfterShutdown() {
|
||||||
|
when(executorService.isShutdown()).thenReturn(true);
|
||||||
|
getRecordsCache.getNextResult();
|
||||||
|
}
|
||||||
|
|
||||||
|
@After
|
||||||
|
public void shutdown() {
|
||||||
|
getRecordsCache.shutdown();
|
||||||
|
verify(executorService).shutdownNow();
|
||||||
|
}
|
||||||
|
|
||||||
|
private void sleep(long millis) {
|
||||||
|
try {
|
||||||
|
Thread.sleep(millis);
|
||||||
|
} catch (InterruptedException e) {}
|
||||||
|
}
|
||||||
|
|
||||||
|
private ByteBuffer createByteBufferWithSize(int size) {
|
||||||
|
ByteBuffer byteBuffer = ByteBuffer.allocate(size);
|
||||||
|
byteBuffer.put(new byte[size]);
|
||||||
|
return byteBuffer;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
@ -18,8 +18,7 @@ import static org.junit.Assert.assertEquals;
|
||||||
import static org.junit.Assert.assertNull;
|
import static org.junit.Assert.assertNull;
|
||||||
import static org.junit.Assert.assertSame;
|
import static org.junit.Assert.assertSame;
|
||||||
import static org.junit.Assert.assertTrue;
|
import static org.junit.Assert.assertTrue;
|
||||||
import static org.mockito.Matchers.anyInt;
|
import static org.mockito.Matchers.any;
|
||||||
import static org.mockito.Matchers.eq;
|
|
||||||
import static org.mockito.Mockito.doReturn;
|
import static org.mockito.Mockito.doReturn;
|
||||||
import static org.mockito.Mockito.doThrow;
|
import static org.mockito.Mockito.doThrow;
|
||||||
import static org.mockito.Mockito.never;
|
import static org.mockito.Mockito.never;
|
||||||
|
|
@ -49,7 +48,6 @@ import com.amazonaws.services.kinesis.clientlibrary.types.ExtendedSequenceNumber
|
||||||
import com.amazonaws.services.kinesis.clientlibrary.types.Messages.AggregatedRecord;
|
import com.amazonaws.services.kinesis.clientlibrary.types.Messages.AggregatedRecord;
|
||||||
import com.amazonaws.services.kinesis.clientlibrary.types.ProcessRecordsInput;
|
import com.amazonaws.services.kinesis.clientlibrary.types.ProcessRecordsInput;
|
||||||
import com.amazonaws.services.kinesis.clientlibrary.types.UserRecord;
|
import com.amazonaws.services.kinesis.clientlibrary.types.UserRecord;
|
||||||
import com.amazonaws.services.kinesis.model.GetRecordsResult;
|
|
||||||
import com.amazonaws.services.kinesis.model.ProvisionedThroughputExceededException;
|
import com.amazonaws.services.kinesis.model.ProvisionedThroughputExceededException;
|
||||||
import com.amazonaws.services.kinesis.model.Record;
|
import com.amazonaws.services.kinesis.model.Record;
|
||||||
import com.google.protobuf.ByteString;
|
import com.google.protobuf.ByteString;
|
||||||
|
|
@ -77,7 +75,7 @@ public class ProcessTaskTest {
|
||||||
@Mock
|
@Mock
|
||||||
private ThrottlingReporter throttlingReporter;
|
private ThrottlingReporter throttlingReporter;
|
||||||
@Mock
|
@Mock
|
||||||
private GetRecordsRetrievalStrategy mockGetRecordsRetrievalStrategy;
|
private GetRecordsCache getRecordsCache;
|
||||||
|
|
||||||
private List<Record> processedRecords;
|
private List<Record> processedRecords;
|
||||||
private ExtendedSequenceNumber newLargestPermittedCheckpointValue;
|
private ExtendedSequenceNumber newLargestPermittedCheckpointValue;
|
||||||
|
|
@ -95,32 +93,39 @@ public class ProcessTaskTest {
|
||||||
INITIAL_POSITION_LATEST);
|
INITIAL_POSITION_LATEST);
|
||||||
final ShardInfo shardInfo = new ShardInfo(shardId, null, null, null);
|
final ShardInfo shardInfo = new ShardInfo(shardId, null, null, null);
|
||||||
processTask = new ProcessTask(
|
processTask = new ProcessTask(
|
||||||
shardInfo, config, mockRecordProcessor, mockCheckpointer, mockDataFetcher, taskBackoffTimeMillis,
|
shardInfo,
|
||||||
KinesisClientLibConfiguration.DEFAULT_SKIP_SHARD_SYNC_AT_STARTUP_IF_LEASES_EXIST, throttlingReporter, mockGetRecordsRetrievalStrategy);
|
config,
|
||||||
|
mockRecordProcessor,
|
||||||
|
mockCheckpointer,
|
||||||
|
mockDataFetcher,
|
||||||
|
taskBackoffTimeMillis,
|
||||||
|
KinesisClientLibConfiguration.DEFAULT_SKIP_SHARD_SYNC_AT_STARTUP_IF_LEASES_EXIST,
|
||||||
|
throttlingReporter,
|
||||||
|
getRecordsCache);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
public void testProcessTaskWithProvisionedThroughputExceededException() {
|
public void testProcessTaskWithProvisionedThroughputExceededException() {
|
||||||
// Set data fetcher to throw exception
|
// Set data fetcher to throw exception
|
||||||
doReturn(false).when(mockDataFetcher).isShardEndReached();
|
doReturn(false).when(mockDataFetcher).isShardEndReached();
|
||||||
doThrow(new ProvisionedThroughputExceededException("Test Exception")).when(mockGetRecordsRetrievalStrategy)
|
doThrow(new ProvisionedThroughputExceededException("Test Exception")).when(getRecordsCache)
|
||||||
.getRecords(maxRecords);
|
.getNextResult();
|
||||||
|
|
||||||
TaskResult result = processTask.call();
|
TaskResult result = processTask.call();
|
||||||
verify(throttlingReporter).throttled();
|
verify(throttlingReporter).throttled();
|
||||||
verify(throttlingReporter, never()).success();
|
verify(throttlingReporter, never()).success();
|
||||||
verify(mockGetRecordsRetrievalStrategy).getRecords(eq(maxRecords));
|
verify(getRecordsCache).getNextResult();
|
||||||
assertTrue("Result should contain ProvisionedThroughputExceededException",
|
assertTrue("Result should contain ProvisionedThroughputExceededException",
|
||||||
result.getException() instanceof ProvisionedThroughputExceededException);
|
result.getException() instanceof ProvisionedThroughputExceededException);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
public void testProcessTaskWithNonExistentStream() {
|
public void testProcessTaskWithNonExistentStream() {
|
||||||
// Data fetcher returns a null Result when the stream does not exist
|
// Data fetcher returns a null Result ` the stream does not exist
|
||||||
doReturn(null).when(mockGetRecordsRetrievalStrategy).getRecords(maxRecords);
|
doReturn(new ProcessRecordsInput().withRecords(Collections.emptyList()).withMillisBehindLatest((long) 0)).when(getRecordsCache).getNextResult();
|
||||||
|
|
||||||
TaskResult result = processTask.call();
|
TaskResult result = processTask.call();
|
||||||
verify(mockGetRecordsRetrievalStrategy).getRecords(eq(maxRecords));
|
verify(getRecordsCache).getNextResult();
|
||||||
assertNull("Task should not throw an exception", result.getException());
|
assertNull("Task should not throw an exception", result.getException());
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -304,14 +309,13 @@ public class ProcessTaskTest {
|
||||||
private void testWithRecords(List<Record> records,
|
private void testWithRecords(List<Record> records,
|
||||||
ExtendedSequenceNumber lastCheckpointValue,
|
ExtendedSequenceNumber lastCheckpointValue,
|
||||||
ExtendedSequenceNumber largestPermittedCheckpointValue) {
|
ExtendedSequenceNumber largestPermittedCheckpointValue) {
|
||||||
when(mockGetRecordsRetrievalStrategy.getRecords(anyInt())).thenReturn(
|
when(getRecordsCache.getNextResult()).thenReturn(new ProcessRecordsInput().withRecords(records).withMillisBehindLatest((long) 1000 * 50));
|
||||||
new GetRecordsResult().withRecords(records));
|
|
||||||
when(mockCheckpointer.getLastCheckpointValue()).thenReturn(lastCheckpointValue);
|
when(mockCheckpointer.getLastCheckpointValue()).thenReturn(lastCheckpointValue);
|
||||||
when(mockCheckpointer.getLargestPermittedCheckpointValue()).thenReturn(largestPermittedCheckpointValue);
|
when(mockCheckpointer.getLargestPermittedCheckpointValue()).thenReturn(largestPermittedCheckpointValue);
|
||||||
processTask.call();
|
processTask.call();
|
||||||
verify(throttlingReporter).success();
|
verify(throttlingReporter).success();
|
||||||
verify(throttlingReporter, never()).throttled();
|
verify(throttlingReporter, never()).throttled();
|
||||||
verify(mockGetRecordsRetrievalStrategy).getRecords(anyInt());
|
verify(getRecordsCache).getNextResult();
|
||||||
ArgumentCaptor<ProcessRecordsInput> priCaptor = ArgumentCaptor.forClass(ProcessRecordsInput.class);
|
ArgumentCaptor<ProcessRecordsInput> priCaptor = ArgumentCaptor.forClass(ProcessRecordsInput.class);
|
||||||
verify(mockRecordProcessor).processRecords(priCaptor.capture());
|
verify(mockRecordProcessor).processRecords(priCaptor.capture());
|
||||||
processedRecords = priCaptor.getValue().getRecords();
|
processedRecords = priCaptor.getValue().getRecords();
|
||||||
|
|
|
||||||
|
|
@ -0,0 +1,44 @@
|
||||||
|
package com.amazonaws.services.kinesis.clientlibrary.lib.worker;
|
||||||
|
|
||||||
|
import com.amazonaws.services.kinesis.metrics.interfaces.IMetricsFactory;
|
||||||
|
|
||||||
|
import static org.hamcrest.CoreMatchers.instanceOf;
|
||||||
|
import static org.hamcrest.MatcherAssert.assertThat;
|
||||||
|
|
||||||
|
import org.junit.Before;
|
||||||
|
import org.junit.Test;
|
||||||
|
import org.mockito.Mock;
|
||||||
|
import org.mockito.MockitoAnnotations;
|
||||||
|
|
||||||
|
public class RecordsFetcherFactoryTest {
|
||||||
|
private String shardId = "TestShard";
|
||||||
|
private RecordsFetcherFactory recordsFetcherFactory;
|
||||||
|
|
||||||
|
@Mock
|
||||||
|
private GetRecordsRetrievalStrategy getRecordsRetrievalStrategy;
|
||||||
|
|
||||||
|
@Mock
|
||||||
|
private IMetricsFactory metricsFactory;
|
||||||
|
|
||||||
|
@Before
|
||||||
|
public void setUp() {
|
||||||
|
MockitoAnnotations.initMocks(this);
|
||||||
|
recordsFetcherFactory = new SimpleRecordsFetcherFactory(1);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void createDefaultRecordsFetcherTest() {
|
||||||
|
GetRecordsCache recordsCache = recordsFetcherFactory.createRecordsFetcher(getRecordsRetrievalStrategy, shardId,
|
||||||
|
metricsFactory);
|
||||||
|
assertThat(recordsCache, instanceOf(BlockingGetRecordsCache.class));
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void createPrefetchRecordsFetcherTest() {
|
||||||
|
recordsFetcherFactory.setDataFetchingStrategy(DataFetchingStrategy.PREFETCH_CACHED);
|
||||||
|
GetRecordsCache recordsCache = recordsFetcherFactory.createRecordsFetcher(getRecordsRetrievalStrategy, shardId,
|
||||||
|
metricsFactory);
|
||||||
|
assertThat(recordsCache, instanceOf(PrefetchGetRecordsCache.class));
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
@ -20,7 +20,6 @@ import static org.hamcrest.Matchers.is;
|
||||||
import static org.hamcrest.Matchers.nullValue;
|
import static org.hamcrest.Matchers.nullValue;
|
||||||
import static org.junit.Assert.assertEquals;
|
import static org.junit.Assert.assertEquals;
|
||||||
import static org.junit.Assert.assertThat;
|
import static org.junit.Assert.assertThat;
|
||||||
import static org.junit.Assert.assertTrue;
|
|
||||||
import static org.junit.Assert.fail;
|
import static org.junit.Assert.fail;
|
||||||
import static org.mockito.Matchers.any;
|
import static org.mockito.Matchers.any;
|
||||||
import static org.mockito.Matchers.anyString;
|
import static org.mockito.Matchers.anyString;
|
||||||
|
|
@ -52,6 +51,7 @@ import org.apache.commons.logging.LogFactory;
|
||||||
import org.hamcrest.Description;
|
import org.hamcrest.Description;
|
||||||
import org.hamcrest.Matcher;
|
import org.hamcrest.Matcher;
|
||||||
import org.hamcrest.TypeSafeMatcher;
|
import org.hamcrest.TypeSafeMatcher;
|
||||||
|
import org.junit.Before;
|
||||||
import org.junit.Test;
|
import org.junit.Test;
|
||||||
import org.junit.runner.RunWith;
|
import org.junit.runner.RunWith;
|
||||||
import org.mockito.Mock;
|
import org.mockito.Mock;
|
||||||
|
|
@ -95,10 +95,16 @@ public class ShardConsumerTest {
|
||||||
// Use Executors.newFixedThreadPool since it returns ThreadPoolExecutor, which is
|
// Use Executors.newFixedThreadPool since it returns ThreadPoolExecutor, which is
|
||||||
// ... a non-final public class, and so can be mocked and spied.
|
// ... a non-final public class, and so can be mocked and spied.
|
||||||
private final ExecutorService executorService = Executors.newFixedThreadPool(1);
|
private final ExecutorService executorService = Executors.newFixedThreadPool(1);
|
||||||
|
private final int maxRecords = 500;
|
||||||
|
private RecordsFetcherFactory recordsFetcherFactory;
|
||||||
|
|
||||||
|
private GetRecordsCache getRecordsCache;
|
||||||
|
|
||||||
@Mock
|
@Mock
|
||||||
private IRecordProcessor processor;
|
private IRecordProcessor processor;
|
||||||
@Mock
|
@Mock
|
||||||
|
private KinesisClientLibConfiguration config;
|
||||||
|
@Mock
|
||||||
private IKinesisProxy streamProxy;
|
private IKinesisProxy streamProxy;
|
||||||
@Mock
|
@Mock
|
||||||
private ILeaseManager<KinesisClientLease> leaseManager;
|
private ILeaseManager<KinesisClientLease> leaseManager;
|
||||||
|
|
@ -107,6 +113,14 @@ public class ShardConsumerTest {
|
||||||
@Mock
|
@Mock
|
||||||
private ShutdownNotification shutdownNotification;
|
private ShutdownNotification shutdownNotification;
|
||||||
|
|
||||||
|
@Before
|
||||||
|
public void setup() {
|
||||||
|
getRecordsCache = null;
|
||||||
|
|
||||||
|
recordsFetcherFactory = spy(new SimpleRecordsFetcherFactory(maxRecords));
|
||||||
|
when(config.getRecordsFetcherFactory()).thenReturn(recordsFetcherFactory);
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Test method to verify consumer stays in INITIALIZING state when InitializationTask fails.
|
* Test method to verify consumer stays in INITIALIZING state when InitializationTask fails.
|
||||||
*/
|
*/
|
||||||
|
|
@ -137,7 +151,8 @@ public class ShardConsumerTest {
|
||||||
executorService,
|
executorService,
|
||||||
metricsFactory,
|
metricsFactory,
|
||||||
taskBackoffTimeMillis,
|
taskBackoffTimeMillis,
|
||||||
KinesisClientLibConfiguration.DEFAULT_SKIP_SHARD_SYNC_AT_STARTUP_IF_LEASES_EXIST);
|
KinesisClientLibConfiguration.DEFAULT_SKIP_SHARD_SYNC_AT_STARTUP_IF_LEASES_EXIST,
|
||||||
|
config);
|
||||||
|
|
||||||
assertThat(consumer.getCurrentState(), is(equalTo(ConsumerStates.ShardConsumerState.WAITING_ON_PARENT_SHARDS)));
|
assertThat(consumer.getCurrentState(), is(equalTo(ConsumerStates.ShardConsumerState.WAITING_ON_PARENT_SHARDS)));
|
||||||
consumer.consumeShard(); // initialize
|
consumer.consumeShard(); // initialize
|
||||||
|
|
@ -154,7 +169,6 @@ public class ShardConsumerTest {
|
||||||
assertThat(consumer.getCurrentState(), is(equalTo(ConsumerStates.ShardConsumerState.INITIALIZING)));
|
assertThat(consumer.getCurrentState(), is(equalTo(ConsumerStates.ShardConsumerState.INITIALIZING)));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Test method to verify consumer stays in INITIALIZING state when InitializationTask fails.
|
* Test method to verify consumer stays in INITIALIZING state when InitializationTask fails.
|
||||||
*/
|
*/
|
||||||
|
|
@ -185,7 +199,8 @@ public class ShardConsumerTest {
|
||||||
spyExecutorService,
|
spyExecutorService,
|
||||||
metricsFactory,
|
metricsFactory,
|
||||||
taskBackoffTimeMillis,
|
taskBackoffTimeMillis,
|
||||||
KinesisClientLibConfiguration.DEFAULT_SKIP_SHARD_SYNC_AT_STARTUP_IF_LEASES_EXIST);
|
KinesisClientLibConfiguration.DEFAULT_SKIP_SHARD_SYNC_AT_STARTUP_IF_LEASES_EXIST,
|
||||||
|
config);
|
||||||
|
|
||||||
assertThat(consumer.getCurrentState(), is(equalTo(ConsumerStates.ShardConsumerState.WAITING_ON_PARENT_SHARDS)));
|
assertThat(consumer.getCurrentState(), is(equalTo(ConsumerStates.ShardConsumerState.WAITING_ON_PARENT_SHARDS)));
|
||||||
consumer.consumeShard(); // initialize
|
consumer.consumeShard(); // initialize
|
||||||
|
|
@ -226,7 +241,8 @@ public class ShardConsumerTest {
|
||||||
executorService,
|
executorService,
|
||||||
metricsFactory,
|
metricsFactory,
|
||||||
taskBackoffTimeMillis,
|
taskBackoffTimeMillis,
|
||||||
KinesisClientLibConfiguration.DEFAULT_SKIP_SHARD_SYNC_AT_STARTUP_IF_LEASES_EXIST);
|
KinesisClientLibConfiguration.DEFAULT_SKIP_SHARD_SYNC_AT_STARTUP_IF_LEASES_EXIST,
|
||||||
|
config);
|
||||||
|
|
||||||
final ExtendedSequenceNumber checkpointSequenceNumber = new ExtendedSequenceNumber("123");
|
final ExtendedSequenceNumber checkpointSequenceNumber = new ExtendedSequenceNumber("123");
|
||||||
final ExtendedSequenceNumber pendingCheckpointSequenceNumber = null;
|
final ExtendedSequenceNumber pendingCheckpointSequenceNumber = null;
|
||||||
|
|
@ -299,7 +315,6 @@ public class ShardConsumerTest {
|
||||||
ICheckpoint checkpoint = new InMemoryCheckpointImpl(startSeqNum.toString());
|
ICheckpoint checkpoint = new InMemoryCheckpointImpl(startSeqNum.toString());
|
||||||
checkpoint.setCheckpoint(streamShardId, ExtendedSequenceNumber.TRIM_HORIZON, testConcurrencyToken);
|
checkpoint.setCheckpoint(streamShardId, ExtendedSequenceNumber.TRIM_HORIZON, testConcurrencyToken);
|
||||||
when(leaseManager.getLease(anyString())).thenReturn(null);
|
when(leaseManager.getLease(anyString())).thenReturn(null);
|
||||||
|
|
||||||
TestStreamlet processor = new TestStreamlet();
|
TestStreamlet processor = new TestStreamlet();
|
||||||
|
|
||||||
StreamConfig streamConfig =
|
StreamConfig streamConfig =
|
||||||
|
|
@ -310,18 +325,41 @@ public class ShardConsumerTest {
|
||||||
skipCheckpointValidationValue, INITIAL_POSITION_LATEST);
|
skipCheckpointValidationValue, INITIAL_POSITION_LATEST);
|
||||||
|
|
||||||
ShardInfo shardInfo = new ShardInfo(streamShardId, testConcurrencyToken, null, null);
|
ShardInfo shardInfo = new ShardInfo(streamShardId, testConcurrencyToken, null, null);
|
||||||
|
|
||||||
|
RecordProcessorCheckpointer recordProcessorCheckpointer = new RecordProcessorCheckpointer(
|
||||||
|
shardInfo,
|
||||||
|
checkpoint,
|
||||||
|
new SequenceNumberValidator(
|
||||||
|
streamConfig.getStreamProxy(),
|
||||||
|
shardInfo.getShardId(),
|
||||||
|
streamConfig.shouldValidateSequenceNumberBeforeCheckpointing()
|
||||||
|
)
|
||||||
|
);
|
||||||
|
|
||||||
|
KinesisDataFetcher dataFetcher = new KinesisDataFetcher(streamConfig.getStreamProxy(), shardInfo);
|
||||||
|
|
||||||
|
getRecordsCache = spy(new BlockingGetRecordsCache(maxRecords,
|
||||||
|
new SynchronousGetRecordsRetrievalStrategy(dataFetcher),
|
||||||
|
0L));
|
||||||
|
when(recordsFetcherFactory.createRecordsFetcher(any(), anyString(),any())).thenReturn(getRecordsCache);
|
||||||
|
|
||||||
ShardConsumer consumer =
|
ShardConsumer consumer =
|
||||||
new ShardConsumer(shardInfo,
|
new ShardConsumer(shardInfo,
|
||||||
streamConfig,
|
streamConfig,
|
||||||
checkpoint,
|
checkpoint,
|
||||||
processor,
|
processor,
|
||||||
|
recordProcessorCheckpointer,
|
||||||
leaseManager,
|
leaseManager,
|
||||||
parentShardPollIntervalMillis,
|
parentShardPollIntervalMillis,
|
||||||
cleanupLeasesOfCompletedShards,
|
cleanupLeasesOfCompletedShards,
|
||||||
executorService,
|
executorService,
|
||||||
metricsFactory,
|
metricsFactory,
|
||||||
taskBackoffTimeMillis,
|
taskBackoffTimeMillis,
|
||||||
KinesisClientLibConfiguration.DEFAULT_SKIP_SHARD_SYNC_AT_STARTUP_IF_LEASES_EXIST);
|
KinesisClientLibConfiguration.DEFAULT_SKIP_SHARD_SYNC_AT_STARTUP_IF_LEASES_EXIST,
|
||||||
|
dataFetcher,
|
||||||
|
Optional.empty(),
|
||||||
|
Optional.empty(),
|
||||||
|
config);
|
||||||
|
|
||||||
assertThat(consumer.getCurrentState(), is(equalTo(ConsumerStates.ShardConsumerState.WAITING_ON_PARENT_SHARDS)));
|
assertThat(consumer.getCurrentState(), is(equalTo(ConsumerStates.ShardConsumerState.WAITING_ON_PARENT_SHARDS)));
|
||||||
consumer.consumeShard(); // check on parent shards
|
consumer.consumeShard(); // check on parent shards
|
||||||
|
|
@ -330,6 +368,7 @@ public class ShardConsumerTest {
|
||||||
assertThat(consumer.getCurrentState(), is(equalTo(ConsumerStates.ShardConsumerState.INITIALIZING)));
|
assertThat(consumer.getCurrentState(), is(equalTo(ConsumerStates.ShardConsumerState.INITIALIZING)));
|
||||||
consumer.consumeShard(); // initialize
|
consumer.consumeShard(); // initialize
|
||||||
processor.getInitializeLatch().await(5, TimeUnit.SECONDS);
|
processor.getInitializeLatch().await(5, TimeUnit.SECONDS);
|
||||||
|
verify(getRecordsCache).start();
|
||||||
|
|
||||||
// We expect to process all records in numRecs calls
|
// We expect to process all records in numRecs calls
|
||||||
for (int i = 0; i < numRecs;) {
|
for (int i = 0; i < numRecs;) {
|
||||||
|
|
@ -343,6 +382,8 @@ public class ShardConsumerTest {
|
||||||
Thread.sleep(50L);
|
Thread.sleep(50L);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
verify(getRecordsCache, times(5)).getNextResult();
|
||||||
|
|
||||||
assertThat(processor.getShutdownReason(), nullValue());
|
assertThat(processor.getShutdownReason(), nullValue());
|
||||||
consumer.notifyShutdownRequested(shutdownNotification);
|
consumer.notifyShutdownRequested(shutdownNotification);
|
||||||
consumer.consumeShard();
|
consumer.consumeShard();
|
||||||
|
|
@ -366,6 +407,8 @@ public class ShardConsumerTest {
|
||||||
assertThat(consumer.getCurrentState(), is(equalTo(ConsumerStates.ShardConsumerState.SHUTDOWN_COMPLETE)));
|
assertThat(consumer.getCurrentState(), is(equalTo(ConsumerStates.ShardConsumerState.SHUTDOWN_COMPLETE)));
|
||||||
assertThat(processor.getShutdownReason(), is(equalTo(ShutdownReason.ZOMBIE)));
|
assertThat(processor.getShutdownReason(), is(equalTo(ShutdownReason.ZOMBIE)));
|
||||||
|
|
||||||
|
verify(getRecordsCache).shutdown();
|
||||||
|
|
||||||
executorService.shutdown();
|
executorService.shutdown();
|
||||||
executorService.awaitTermination(60, TimeUnit.SECONDS);
|
executorService.awaitTermination(60, TimeUnit.SECONDS);
|
||||||
|
|
||||||
|
|
@ -401,7 +444,6 @@ public class ShardConsumerTest {
|
||||||
ICheckpoint checkpoint = new InMemoryCheckpointImpl(startSeqNum.toString());
|
ICheckpoint checkpoint = new InMemoryCheckpointImpl(startSeqNum.toString());
|
||||||
checkpoint.setCheckpoint(streamShardId, ExtendedSequenceNumber.AT_TIMESTAMP, testConcurrencyToken);
|
checkpoint.setCheckpoint(streamShardId, ExtendedSequenceNumber.AT_TIMESTAMP, testConcurrencyToken);
|
||||||
when(leaseManager.getLease(anyString())).thenReturn(null);
|
when(leaseManager.getLease(anyString())).thenReturn(null);
|
||||||
|
|
||||||
TestStreamlet processor = new TestStreamlet();
|
TestStreamlet processor = new TestStreamlet();
|
||||||
|
|
||||||
StreamConfig streamConfig =
|
StreamConfig streamConfig =
|
||||||
|
|
@ -413,18 +455,41 @@ public class ShardConsumerTest {
|
||||||
atTimestamp);
|
atTimestamp);
|
||||||
|
|
||||||
ShardInfo shardInfo = new ShardInfo(streamShardId, testConcurrencyToken, null, ExtendedSequenceNumber.TRIM_HORIZON);
|
ShardInfo shardInfo = new ShardInfo(streamShardId, testConcurrencyToken, null, ExtendedSequenceNumber.TRIM_HORIZON);
|
||||||
|
|
||||||
|
RecordProcessorCheckpointer recordProcessorCheckpointer = new RecordProcessorCheckpointer(
|
||||||
|
shardInfo,
|
||||||
|
checkpoint,
|
||||||
|
new SequenceNumberValidator(
|
||||||
|
streamConfig.getStreamProxy(),
|
||||||
|
shardInfo.getShardId(),
|
||||||
|
streamConfig.shouldValidateSequenceNumberBeforeCheckpointing()
|
||||||
|
)
|
||||||
|
);
|
||||||
|
|
||||||
|
KinesisDataFetcher dataFetcher = new KinesisDataFetcher(streamConfig.getStreamProxy(), shardInfo);
|
||||||
|
|
||||||
|
getRecordsCache = spy(new BlockingGetRecordsCache(maxRecords,
|
||||||
|
new SynchronousGetRecordsRetrievalStrategy(dataFetcher),
|
||||||
|
0L));
|
||||||
|
when(recordsFetcherFactory.createRecordsFetcher(any(), anyString(),any())).thenReturn(getRecordsCache);
|
||||||
|
|
||||||
ShardConsumer consumer =
|
ShardConsumer consumer =
|
||||||
new ShardConsumer(shardInfo,
|
new ShardConsumer(shardInfo,
|
||||||
streamConfig,
|
streamConfig,
|
||||||
checkpoint,
|
checkpoint,
|
||||||
processor,
|
processor,
|
||||||
|
recordProcessorCheckpointer,
|
||||||
leaseManager,
|
leaseManager,
|
||||||
parentShardPollIntervalMillis,
|
parentShardPollIntervalMillis,
|
||||||
cleanupLeasesOfCompletedShards,
|
cleanupLeasesOfCompletedShards,
|
||||||
executorService,
|
executorService,
|
||||||
metricsFactory,
|
metricsFactory,
|
||||||
taskBackoffTimeMillis,
|
taskBackoffTimeMillis,
|
||||||
KinesisClientLibConfiguration.DEFAULT_SKIP_SHARD_SYNC_AT_STARTUP_IF_LEASES_EXIST);
|
KinesisClientLibConfiguration.DEFAULT_SKIP_SHARD_SYNC_AT_STARTUP_IF_LEASES_EXIST,
|
||||||
|
dataFetcher,
|
||||||
|
Optional.empty(),
|
||||||
|
Optional.empty(),
|
||||||
|
config);
|
||||||
|
|
||||||
assertThat(consumer.getCurrentState(), is(equalTo(ConsumerStates.ShardConsumerState.WAITING_ON_PARENT_SHARDS)));
|
assertThat(consumer.getCurrentState(), is(equalTo(ConsumerStates.ShardConsumerState.WAITING_ON_PARENT_SHARDS)));
|
||||||
consumer.consumeShard(); // check on parent shards
|
consumer.consumeShard(); // check on parent shards
|
||||||
|
|
@ -434,6 +499,8 @@ public class ShardConsumerTest {
|
||||||
consumer.consumeShard(); // initialize
|
consumer.consumeShard(); // initialize
|
||||||
Thread.sleep(50L);
|
Thread.sleep(50L);
|
||||||
|
|
||||||
|
verify(getRecordsCache).start();
|
||||||
|
|
||||||
// We expect to process all records in numRecs calls
|
// We expect to process all records in numRecs calls
|
||||||
for (int i = 0; i < numRecs;) {
|
for (int i = 0; i < numRecs;) {
|
||||||
boolean newTaskSubmitted = consumer.consumeShard();
|
boolean newTaskSubmitted = consumer.consumeShard();
|
||||||
|
|
@ -446,6 +513,8 @@ public class ShardConsumerTest {
|
||||||
Thread.sleep(50L);
|
Thread.sleep(50L);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
verify(getRecordsCache, times(4)).getNextResult();
|
||||||
|
|
||||||
assertThat(processor.getShutdownReason(), nullValue());
|
assertThat(processor.getShutdownReason(), nullValue());
|
||||||
consumer.beginShutdown();
|
consumer.beginShutdown();
|
||||||
Thread.sleep(50L);
|
Thread.sleep(50L);
|
||||||
|
|
@ -457,8 +526,11 @@ public class ShardConsumerTest {
|
||||||
executorService.shutdown();
|
executorService.shutdown();
|
||||||
executorService.awaitTermination(60, TimeUnit.SECONDS);
|
executorService.awaitTermination(60, TimeUnit.SECONDS);
|
||||||
|
|
||||||
|
verify(getRecordsCache).shutdown();
|
||||||
|
|
||||||
String iterator = fileBasedProxy.getIterator(streamShardId, timestamp);
|
String iterator = fileBasedProxy.getIterator(streamShardId, timestamp);
|
||||||
List<Record> expectedRecords = toUserRecords(fileBasedProxy.get(iterator, numRecs).getRecords());
|
List<Record> expectedRecords = toUserRecords(fileBasedProxy.get(iterator, numRecs).getRecords());
|
||||||
|
|
||||||
verifyConsumedRecords(expectedRecords, processor.getProcessedRecords());
|
verifyConsumedRecords(expectedRecords, processor.getProcessedRecords());
|
||||||
assertEquals(4, processor.getProcessedRecords().size());
|
assertEquals(4, processor.getProcessedRecords().size());
|
||||||
file.delete();
|
file.delete();
|
||||||
|
|
@ -486,11 +558,15 @@ public class ShardConsumerTest {
|
||||||
executorService,
|
executorService,
|
||||||
metricsFactory,
|
metricsFactory,
|
||||||
taskBackoffTimeMillis,
|
taskBackoffTimeMillis,
|
||||||
KinesisClientLibConfiguration.DEFAULT_SKIP_SHARD_SYNC_AT_STARTUP_IF_LEASES_EXIST);
|
KinesisClientLibConfiguration.DEFAULT_SKIP_SHARD_SYNC_AT_STARTUP_IF_LEASES_EXIST,
|
||||||
|
config);
|
||||||
|
|
||||||
|
GetRecordsCache getRecordsCache = spy(consumer.getGetRecordsCache());
|
||||||
|
|
||||||
final ExtendedSequenceNumber checkpointSequenceNumber = new ExtendedSequenceNumber("123");
|
final ExtendedSequenceNumber checkpointSequenceNumber = new ExtendedSequenceNumber("123");
|
||||||
final ExtendedSequenceNumber pendingCheckpointSequenceNumber = new ExtendedSequenceNumber("999");
|
final ExtendedSequenceNumber pendingCheckpointSequenceNumber = new ExtendedSequenceNumber("999");
|
||||||
when(leaseManager.getLease(anyString())).thenReturn(null);
|
when(leaseManager.getLease(anyString())).thenReturn(null);
|
||||||
|
when(config.getRecordsFetcherFactory()).thenReturn(new SimpleRecordsFetcherFactory(2));
|
||||||
when(checkpoint.getCheckpointObject(anyString())).thenReturn(
|
when(checkpoint.getCheckpointObject(anyString())).thenReturn(
|
||||||
new Checkpoint(checkpointSequenceNumber, pendingCheckpointSequenceNumber));
|
new Checkpoint(checkpointSequenceNumber, pendingCheckpointSequenceNumber));
|
||||||
|
|
||||||
|
|
@ -535,9 +611,11 @@ public class ShardConsumerTest {
|
||||||
taskBackoffTimeMillis,
|
taskBackoffTimeMillis,
|
||||||
KinesisClientLibConfiguration.DEFAULT_SKIP_SHARD_SYNC_AT_STARTUP_IF_LEASES_EXIST,
|
KinesisClientLibConfiguration.DEFAULT_SKIP_SHARD_SYNC_AT_STARTUP_IF_LEASES_EXIST,
|
||||||
Optional.empty(),
|
Optional.empty(),
|
||||||
Optional.empty());
|
Optional.empty(),
|
||||||
|
config);
|
||||||
|
|
||||||
assertEquals(shardConsumer.getGetRecordsRetrievalStrategy().getClass(), SynchronousGetRecordsRetrievalStrategy.class);
|
assertEquals(shardConsumer.getGetRecordsCache().getGetRecordsRetrievalStrategy().getClass(),
|
||||||
|
SynchronousGetRecordsRetrievalStrategy.class);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
|
|
@ -563,9 +641,11 @@ public class ShardConsumerTest {
|
||||||
taskBackoffTimeMillis,
|
taskBackoffTimeMillis,
|
||||||
KinesisClientLibConfiguration.DEFAULT_SKIP_SHARD_SYNC_AT_STARTUP_IF_LEASES_EXIST,
|
KinesisClientLibConfiguration.DEFAULT_SKIP_SHARD_SYNC_AT_STARTUP_IF_LEASES_EXIST,
|
||||||
Optional.of(1),
|
Optional.of(1),
|
||||||
Optional.of(2));
|
Optional.of(2),
|
||||||
|
config);
|
||||||
|
|
||||||
assertEquals(shardConsumer.getGetRecordsRetrievalStrategy().getClass(), AsynchronousGetRecordsRetrievalStrategy.class);
|
assertEquals(shardConsumer.getGetRecordsCache().getGetRecordsRetrievalStrategy().getClass(),
|
||||||
|
AsynchronousGetRecordsRetrievalStrategy.class);
|
||||||
}
|
}
|
||||||
|
|
||||||
//@formatter:off (gets the formatting wrong)
|
//@formatter:off (gets the formatting wrong)
|
||||||
|
|
|
||||||
|
|
@ -59,7 +59,7 @@ public class ShutdownTaskTest {
|
||||||
IRecordProcessor defaultRecordProcessor = new TestStreamlet();
|
IRecordProcessor defaultRecordProcessor = new TestStreamlet();
|
||||||
|
|
||||||
@Mock
|
@Mock
|
||||||
private GetRecordsRetrievalStrategy getRecordsRetrievalStrategy;
|
private GetRecordsCache getRecordsCache;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* @throws java.lang.Exception
|
* @throws java.lang.Exception
|
||||||
|
|
@ -80,7 +80,7 @@ public class ShutdownTaskTest {
|
||||||
*/
|
*/
|
||||||
@Before
|
@Before
|
||||||
public void setUp() throws Exception {
|
public void setUp() throws Exception {
|
||||||
doNothing().when(getRecordsRetrievalStrategy).shutdown();
|
doNothing().when(getRecordsCache).shutdown();
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
|
@ -109,7 +109,7 @@ public class ShutdownTaskTest {
|
||||||
cleanupLeasesOfCompletedShards,
|
cleanupLeasesOfCompletedShards,
|
||||||
leaseManager,
|
leaseManager,
|
||||||
TASK_BACKOFF_TIME_MILLIS,
|
TASK_BACKOFF_TIME_MILLIS,
|
||||||
getRecordsRetrievalStrategy);
|
getRecordsCache);
|
||||||
TaskResult result = task.call();
|
TaskResult result = task.call();
|
||||||
Assert.assertNotNull(result.getException());
|
Assert.assertNotNull(result.getException());
|
||||||
Assert.assertTrue(result.getException() instanceof IllegalArgumentException);
|
Assert.assertTrue(result.getException() instanceof IllegalArgumentException);
|
||||||
|
|
@ -135,11 +135,11 @@ public class ShutdownTaskTest {
|
||||||
cleanupLeasesOfCompletedShards,
|
cleanupLeasesOfCompletedShards,
|
||||||
leaseManager,
|
leaseManager,
|
||||||
TASK_BACKOFF_TIME_MILLIS,
|
TASK_BACKOFF_TIME_MILLIS,
|
||||||
getRecordsRetrievalStrategy);
|
getRecordsCache);
|
||||||
TaskResult result = task.call();
|
TaskResult result = task.call();
|
||||||
Assert.assertNotNull(result.getException());
|
Assert.assertNotNull(result.getException());
|
||||||
Assert.assertTrue(result.getException() instanceof KinesisClientLibIOException);
|
Assert.assertTrue(result.getException() instanceof KinesisClientLibIOException);
|
||||||
verify(getRecordsRetrievalStrategy).shutdown();
|
verify(getRecordsCache).shutdown();
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
|
@ -147,7 +147,7 @@ public class ShutdownTaskTest {
|
||||||
*/
|
*/
|
||||||
@Test
|
@Test
|
||||||
public final void testGetTaskType() {
|
public final void testGetTaskType() {
|
||||||
ShutdownTask task = new ShutdownTask(null, null, null, null, null, null, false, null, 0, getRecordsRetrievalStrategy);
|
ShutdownTask task = new ShutdownTask(null, null, null, null, null, null, false, null, 0, getRecordsCache);
|
||||||
Assert.assertEquals(TaskType.SHUTDOWN, task.getTaskType());
|
Assert.assertEquals(TaskType.SHUTDOWN, task.getTaskType());
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -21,10 +21,20 @@ import static org.hamcrest.MatcherAssert.assertThat;
|
||||||
import static org.hamcrest.Matchers.greaterThanOrEqualTo;
|
import static org.hamcrest.Matchers.greaterThanOrEqualTo;
|
||||||
import static org.junit.Assert.fail;
|
import static org.junit.Assert.fail;
|
||||||
import static org.mockito.Matchers.any;
|
import static org.mockito.Matchers.any;
|
||||||
|
import static org.mockito.Matchers.anyString;
|
||||||
import static org.mockito.Matchers.argThat;
|
import static org.mockito.Matchers.argThat;
|
||||||
import static org.mockito.Matchers.eq;
|
import static org.mockito.Matchers.eq;
|
||||||
import static org.mockito.Matchers.same;
|
import static org.mockito.Matchers.same;
|
||||||
import static org.mockito.Mockito.*;
|
import static org.mockito.Mockito.atLeast;
|
||||||
|
import static org.mockito.Mockito.atLeastOnce;
|
||||||
|
import static org.mockito.Mockito.doAnswer;
|
||||||
|
import static org.mockito.Mockito.doReturn;
|
||||||
|
import static org.mockito.Mockito.mock;
|
||||||
|
import static org.mockito.Mockito.never;
|
||||||
|
import static org.mockito.Mockito.spy;
|
||||||
|
import static org.mockito.Mockito.times;
|
||||||
|
import static org.mockito.Mockito.verify;
|
||||||
|
import static org.mockito.Mockito.when;
|
||||||
|
|
||||||
import java.io.File;
|
import java.io.File;
|
||||||
import java.lang.Thread.State;
|
import java.lang.Thread.State;
|
||||||
|
|
@ -60,6 +70,7 @@ import org.hamcrest.Matcher;
|
||||||
import org.hamcrest.TypeSafeDiagnosingMatcher;
|
import org.hamcrest.TypeSafeDiagnosingMatcher;
|
||||||
import org.hamcrest.TypeSafeMatcher;
|
import org.hamcrest.TypeSafeMatcher;
|
||||||
import org.junit.Assert;
|
import org.junit.Assert;
|
||||||
|
import org.junit.Before;
|
||||||
import org.junit.Test;
|
import org.junit.Test;
|
||||||
import org.junit.runner.RunWith;
|
import org.junit.runner.RunWith;
|
||||||
import org.mockito.Matchers;
|
import org.mockito.Matchers;
|
||||||
|
|
@ -130,6 +141,9 @@ public class WorkerTest {
|
||||||
private static final String KINESIS_SHARD_ID_FORMAT = "kinesis-0-0-%d";
|
private static final String KINESIS_SHARD_ID_FORMAT = "kinesis-0-0-%d";
|
||||||
private static final String CONCURRENCY_TOKEN_FORMAT = "testToken-%d";
|
private static final String CONCURRENCY_TOKEN_FORMAT = "testToken-%d";
|
||||||
|
|
||||||
|
private RecordsFetcherFactory recordsFetcherFactory;
|
||||||
|
private KinesisClientLibConfiguration config;
|
||||||
|
|
||||||
@Mock
|
@Mock
|
||||||
private KinesisClientLibLeaseCoordinator leaseCoordinator;
|
private KinesisClientLibLeaseCoordinator leaseCoordinator;
|
||||||
@Mock
|
@Mock
|
||||||
|
|
@ -155,6 +169,13 @@ public class WorkerTest {
|
||||||
@Mock
|
@Mock
|
||||||
private TaskResult taskResult;
|
private TaskResult taskResult;
|
||||||
|
|
||||||
|
@Before
|
||||||
|
public void setup() {
|
||||||
|
config = spy(new KinesisClientLibConfiguration("app", null, null, null));
|
||||||
|
recordsFetcherFactory = spy(new SimpleRecordsFetcherFactory(500));
|
||||||
|
when(config.getRecordsFetcherFactory()).thenReturn(recordsFetcherFactory);
|
||||||
|
}
|
||||||
|
|
||||||
// CHECKSTYLE:IGNORE AnonInnerLengthCheck FOR NEXT 50 LINES
|
// CHECKSTYLE:IGNORE AnonInnerLengthCheck FOR NEXT 50 LINES
|
||||||
private static final com.amazonaws.services.kinesis.clientlibrary.interfaces.IRecordProcessorFactory SAMPLE_RECORD_PROCESSOR_FACTORY =
|
private static final com.amazonaws.services.kinesis.clientlibrary.interfaces.IRecordProcessorFactory SAMPLE_RECORD_PROCESSOR_FACTORY =
|
||||||
new com.amazonaws.services.kinesis.clientlibrary.interfaces.IRecordProcessorFactory() {
|
new com.amazonaws.services.kinesis.clientlibrary.interfaces.IRecordProcessorFactory() {
|
||||||
|
|
@ -195,14 +216,13 @@ public class WorkerTest {
|
||||||
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Test method for {@link com.amazonaws.services.kinesis.clientlibrary.lib.worker.Worker#getApplicationName()}.
|
* Test method for {@link Worker#getApplicationName()}.
|
||||||
*/
|
*/
|
||||||
@Test
|
@Test
|
||||||
public final void testGetStageName() {
|
public final void testGetStageName() {
|
||||||
final String stageName = "testStageName";
|
final String stageName = "testStageName";
|
||||||
final KinesisClientLibConfiguration clientConfig =
|
config = new KinesisClientLibConfiguration(stageName, null, null, null);
|
||||||
new KinesisClientLibConfiguration(stageName, null, null, null);
|
Worker worker = new Worker(v1RecordProcessorFactory, config);
|
||||||
Worker worker = new Worker(v1RecordProcessorFactory, clientConfig);
|
|
||||||
Assert.assertEquals(stageName, worker.getApplicationName());
|
Assert.assertEquals(stageName, worker.getApplicationName());
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -210,6 +230,7 @@ public class WorkerTest {
|
||||||
public final void testCreateOrGetShardConsumer() {
|
public final void testCreateOrGetShardConsumer() {
|
||||||
final String stageName = "testStageName";
|
final String stageName = "testStageName";
|
||||||
IRecordProcessorFactory streamletFactory = SAMPLE_RECORD_PROCESSOR_FACTORY_V2;
|
IRecordProcessorFactory streamletFactory = SAMPLE_RECORD_PROCESSOR_FACTORY_V2;
|
||||||
|
config = new KinesisClientLibConfiguration(stageName, null, null, null);
|
||||||
IKinesisProxy proxy = null;
|
IKinesisProxy proxy = null;
|
||||||
ICheckpoint checkpoint = null;
|
ICheckpoint checkpoint = null;
|
||||||
int maxRecords = 1;
|
int maxRecords = 1;
|
||||||
|
|
@ -228,7 +249,9 @@ public class WorkerTest {
|
||||||
|
|
||||||
Worker worker =
|
Worker worker =
|
||||||
new Worker(stageName,
|
new Worker(stageName,
|
||||||
streamletFactory, streamConfig, INITIAL_POSITION_LATEST,
|
streamletFactory,
|
||||||
|
config,
|
||||||
|
streamConfig, INITIAL_POSITION_LATEST,
|
||||||
parentShardPollIntervalMillis,
|
parentShardPollIntervalMillis,
|
||||||
shardSyncIntervalMillis,
|
shardSyncIntervalMillis,
|
||||||
cleanupLeasesUponShardCompletion,
|
cleanupLeasesUponShardCompletion,
|
||||||
|
|
@ -275,10 +298,22 @@ public class WorkerTest {
|
||||||
when(leaseCoordinator.getCurrentAssignments()).thenReturn(initialState).thenReturn(firstCheckpoint)
|
when(leaseCoordinator.getCurrentAssignments()).thenReturn(initialState).thenReturn(firstCheckpoint)
|
||||||
.thenReturn(secondCheckpoint);
|
.thenReturn(secondCheckpoint);
|
||||||
|
|
||||||
Worker worker = new Worker(stageName, streamletFactory, streamConfig, INITIAL_POSITION_LATEST,
|
Worker worker = new Worker(stageName,
|
||||||
parentShardPollIntervalMillis, shardSyncIntervalMillis, cleanupLeasesUponShardCompletion, checkpoint,
|
streamletFactory,
|
||||||
leaseCoordinator, execService, nullMetricsFactory, taskBackoffTimeMillis, failoverTimeMillis,
|
config,
|
||||||
KinesisClientLibConfiguration.DEFAULT_SKIP_SHARD_SYNC_AT_STARTUP_IF_LEASES_EXIST, shardPrioritization);
|
streamConfig,
|
||||||
|
INITIAL_POSITION_LATEST,
|
||||||
|
parentShardPollIntervalMillis,
|
||||||
|
shardSyncIntervalMillis,
|
||||||
|
cleanupLeasesUponShardCompletion,
|
||||||
|
checkpoint,
|
||||||
|
leaseCoordinator,
|
||||||
|
execService,
|
||||||
|
nullMetricsFactory,
|
||||||
|
taskBackoffTimeMillis,
|
||||||
|
failoverTimeMillis,
|
||||||
|
KinesisClientLibConfiguration.DEFAULT_SKIP_SHARD_SYNC_AT_STARTUP_IF_LEASES_EXIST,
|
||||||
|
shardPrioritization);
|
||||||
|
|
||||||
Worker workerSpy = spy(worker);
|
Worker workerSpy = spy(worker);
|
||||||
|
|
||||||
|
|
@ -314,6 +349,7 @@ public class WorkerTest {
|
||||||
public final void testCleanupShardConsumers() {
|
public final void testCleanupShardConsumers() {
|
||||||
final String stageName = "testStageName";
|
final String stageName = "testStageName";
|
||||||
IRecordProcessorFactory streamletFactory = SAMPLE_RECORD_PROCESSOR_FACTORY_V2;
|
IRecordProcessorFactory streamletFactory = SAMPLE_RECORD_PROCESSOR_FACTORY_V2;
|
||||||
|
config = new KinesisClientLibConfiguration(stageName, null, null, null);
|
||||||
IKinesisProxy proxy = null;
|
IKinesisProxy proxy = null;
|
||||||
ICheckpoint checkpoint = null;
|
ICheckpoint checkpoint = null;
|
||||||
int maxRecords = 1;
|
int maxRecords = 1;
|
||||||
|
|
@ -332,7 +368,9 @@ public class WorkerTest {
|
||||||
|
|
||||||
Worker worker =
|
Worker worker =
|
||||||
new Worker(stageName,
|
new Worker(stageName,
|
||||||
streamletFactory, streamConfig, INITIAL_POSITION_LATEST,
|
streamletFactory,
|
||||||
|
config,
|
||||||
|
streamConfig, INITIAL_POSITION_LATEST,
|
||||||
parentShardPollIntervalMillis,
|
parentShardPollIntervalMillis,
|
||||||
shardSyncIntervalMillis,
|
shardSyncIntervalMillis,
|
||||||
cleanupLeasesUponShardCompletion,
|
cleanupLeasesUponShardCompletion,
|
||||||
|
|
@ -371,6 +409,7 @@ public class WorkerTest {
|
||||||
public final void testInitializationFailureWithRetries() {
|
public final void testInitializationFailureWithRetries() {
|
||||||
String stageName = "testInitializationWorker";
|
String stageName = "testInitializationWorker";
|
||||||
IRecordProcessorFactory recordProcessorFactory = new TestStreamletFactory(null, null);
|
IRecordProcessorFactory recordProcessorFactory = new TestStreamletFactory(null, null);
|
||||||
|
config = new KinesisClientLibConfiguration(stageName, null, null, null);
|
||||||
int count = 0;
|
int count = 0;
|
||||||
when(proxy.getShardList()).thenThrow(new RuntimeException(Integer.toString(count++)));
|
when(proxy.getShardList()).thenThrow(new RuntimeException(Integer.toString(count++)));
|
||||||
int maxRecords = 2;
|
int maxRecords = 2;
|
||||||
|
|
@ -386,6 +425,7 @@ public class WorkerTest {
|
||||||
Worker worker =
|
Worker worker =
|
||||||
new Worker(stageName,
|
new Worker(stageName,
|
||||||
recordProcessorFactory,
|
recordProcessorFactory,
|
||||||
|
config,
|
||||||
streamConfig, INITIAL_POSITION_TRIM_HORIZON,
|
streamConfig, INITIAL_POSITION_TRIM_HORIZON,
|
||||||
shardPollInterval,
|
shardPollInterval,
|
||||||
shardSyncIntervalMillis,
|
shardSyncIntervalMillis,
|
||||||
|
|
@ -437,7 +477,7 @@ public class WorkerTest {
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Runs worker with threadPoolSize < numShards
|
* Runs worker with threadPoolSize < numShards
|
||||||
* Test method for {@link com.amazonaws.services.kinesis.clientlibrary.lib.worker.Worker#run()}.
|
* Test method for {@link Worker#run()}.
|
||||||
*/
|
*/
|
||||||
@Test
|
@Test
|
||||||
public final void testOneSplitShard2Threads() throws Exception {
|
public final void testOneSplitShard2Threads() throws Exception {
|
||||||
|
|
@ -448,12 +488,12 @@ public class WorkerTest {
|
||||||
KinesisClientLease lease = ShardSyncer.newKCLLease(shardList.get(0));
|
KinesisClientLease lease = ShardSyncer.newKCLLease(shardList.get(0));
|
||||||
lease.setCheckpoint(new ExtendedSequenceNumber("2"));
|
lease.setCheckpoint(new ExtendedSequenceNumber("2"));
|
||||||
initialLeases.add(lease);
|
initialLeases.add(lease);
|
||||||
runAndTestWorker(shardList, threadPoolSize, initialLeases, callProcessRecordsForEmptyRecordList, numberOfRecordsPerShard);
|
runAndTestWorker(shardList, threadPoolSize, initialLeases, callProcessRecordsForEmptyRecordList, numberOfRecordsPerShard, config);
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Runs worker with threadPoolSize < numShards
|
* Runs worker with threadPoolSize < numShards
|
||||||
* Test method for {@link com.amazonaws.services.kinesis.clientlibrary.lib.worker.Worker#run()}.
|
* Test method for {@link Worker#run()}.
|
||||||
*/
|
*/
|
||||||
@Test
|
@Test
|
||||||
public final void testOneSplitShard2ThreadsWithCallsForEmptyRecords() throws Exception {
|
public final void testOneSplitShard2ThreadsWithCallsForEmptyRecords() throws Exception {
|
||||||
|
|
@ -465,7 +505,10 @@ public class WorkerTest {
|
||||||
lease.setCheckpoint(new ExtendedSequenceNumber("2"));
|
lease.setCheckpoint(new ExtendedSequenceNumber("2"));
|
||||||
initialLeases.add(lease);
|
initialLeases.add(lease);
|
||||||
boolean callProcessRecordsForEmptyRecordList = true;
|
boolean callProcessRecordsForEmptyRecordList = true;
|
||||||
runAndTestWorker(shardList, threadPoolSize, initialLeases, callProcessRecordsForEmptyRecordList, numberOfRecordsPerShard);
|
RecordsFetcherFactory recordsFetcherFactory = new SimpleRecordsFetcherFactory(500);
|
||||||
|
recordsFetcherFactory.setIdleMillisBetweenCalls(0L);
|
||||||
|
when(config.getRecordsFetcherFactory()).thenReturn(recordsFetcherFactory);
|
||||||
|
runAndTestWorker(shardList, threadPoolSize, initialLeases, callProcessRecordsForEmptyRecordList, numberOfRecordsPerShard, config);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
|
|
@ -490,7 +533,8 @@ public class WorkerTest {
|
||||||
10,
|
10,
|
||||||
kinesisProxy, v2RecordProcessorFactory,
|
kinesisProxy, v2RecordProcessorFactory,
|
||||||
executorService,
|
executorService,
|
||||||
cwMetricsFactory);
|
cwMetricsFactory,
|
||||||
|
config);
|
||||||
|
|
||||||
// Give some time for thread to run.
|
// Give some time for thread to run.
|
||||||
workerStarted.await();
|
workerStarted.await();
|
||||||
|
|
@ -526,7 +570,8 @@ public class WorkerTest {
|
||||||
10,
|
10,
|
||||||
kinesisProxy, v2RecordProcessorFactory,
|
kinesisProxy, v2RecordProcessorFactory,
|
||||||
executorService,
|
executorService,
|
||||||
cwMetricsFactory);
|
cwMetricsFactory,
|
||||||
|
config);
|
||||||
|
|
||||||
// Give some time for thread to run.
|
// Give some time for thread to run.
|
||||||
workerStarted.await();
|
workerStarted.await();
|
||||||
|
|
@ -573,6 +618,12 @@ public class WorkerTest {
|
||||||
}
|
}
|
||||||
}).when(v2RecordProcessor).processRecords(any(ProcessRecordsInput.class));
|
}).when(v2RecordProcessor).processRecords(any(ProcessRecordsInput.class));
|
||||||
|
|
||||||
|
RecordsFetcherFactory recordsFetcherFactory = mock(RecordsFetcherFactory.class);
|
||||||
|
GetRecordsCache getRecordsCache = mock(GetRecordsCache.class);
|
||||||
|
when(config.getRecordsFetcherFactory()).thenReturn(recordsFetcherFactory);
|
||||||
|
when(recordsFetcherFactory.createRecordsFetcher(any(), anyString(),any())).thenReturn(getRecordsCache);
|
||||||
|
when(getRecordsCache.getNextResult()).thenReturn(new ProcessRecordsInput().withRecords(Collections.emptyList()).withMillisBehindLatest(0L));
|
||||||
|
|
||||||
WorkerThread workerThread = runWorker(shardList,
|
WorkerThread workerThread = runWorker(shardList,
|
||||||
initialLeases,
|
initialLeases,
|
||||||
callProcessRecordsForEmptyRecordList,
|
callProcessRecordsForEmptyRecordList,
|
||||||
|
|
@ -581,7 +632,8 @@ public class WorkerTest {
|
||||||
fileBasedProxy,
|
fileBasedProxy,
|
||||||
v2RecordProcessorFactory,
|
v2RecordProcessorFactory,
|
||||||
executorService,
|
executorService,
|
||||||
nullMetricsFactory);
|
nullMetricsFactory,
|
||||||
|
config);
|
||||||
|
|
||||||
// Only sleep for time that is required.
|
// Only sleep for time that is required.
|
||||||
processRecordsLatch.await();
|
processRecordsLatch.await();
|
||||||
|
|
@ -672,7 +724,8 @@ public class WorkerTest {
|
||||||
fileBasedProxy,
|
fileBasedProxy,
|
||||||
v2RecordProcessorFactory,
|
v2RecordProcessorFactory,
|
||||||
executorService,
|
executorService,
|
||||||
nullMetricsFactory);
|
nullMetricsFactory,
|
||||||
|
config);
|
||||||
|
|
||||||
// Only sleep for time that is required.
|
// Only sleep for time that is required.
|
||||||
processRecordsLatch.await();
|
processRecordsLatch.await();
|
||||||
|
|
@ -742,10 +795,22 @@ public class WorkerTest {
|
||||||
when(recordProcessorFactory.createProcessor()).thenReturn(processor);
|
when(recordProcessorFactory.createProcessor()).thenReturn(processor);
|
||||||
|
|
||||||
|
|
||||||
Worker worker = new Worker("testRequestShutdown", recordProcessorFactory, streamConfig,
|
Worker worker = new Worker("testRequestShutdown",
|
||||||
INITIAL_POSITION_TRIM_HORIZON, parentShardPollIntervalMillis, shardSyncIntervalMillis,
|
recordProcessorFactory,
|
||||||
cleanupLeasesUponShardCompletion, leaseCoordinator, leaseCoordinator, executorService, metricsFactory,
|
config,
|
||||||
taskBackoffTimeMillis, failoverTimeMillis, false, shardPrioritization);
|
streamConfig,
|
||||||
|
INITIAL_POSITION_TRIM_HORIZON,
|
||||||
|
parentShardPollIntervalMillis,
|
||||||
|
shardSyncIntervalMillis,
|
||||||
|
cleanupLeasesUponShardCompletion,
|
||||||
|
leaseCoordinator,
|
||||||
|
leaseCoordinator,
|
||||||
|
executorService,
|
||||||
|
metricsFactory,
|
||||||
|
taskBackoffTimeMillis,
|
||||||
|
failoverTimeMillis,
|
||||||
|
false,
|
||||||
|
shardPrioritization);
|
||||||
|
|
||||||
when(executorService.submit(Matchers.<Callable<TaskResult>> any()))
|
when(executorService.submit(Matchers.<Callable<TaskResult>> any()))
|
||||||
.thenAnswer(new ShutdownHandlingAnswer(taskFuture));
|
.thenAnswer(new ShutdownHandlingAnswer(taskFuture));
|
||||||
|
|
@ -816,7 +881,7 @@ public class WorkerTest {
|
||||||
IRecordProcessor processor = mock(IRecordProcessor.class);
|
IRecordProcessor processor = mock(IRecordProcessor.class);
|
||||||
when(recordProcessorFactory.createProcessor()).thenReturn(processor);
|
when(recordProcessorFactory.createProcessor()).thenReturn(processor);
|
||||||
|
|
||||||
Worker worker = new InjectableWorker("testRequestShutdown", recordProcessorFactory, streamConfig,
|
Worker worker = new InjectableWorker("testRequestShutdown", recordProcessorFactory, config, streamConfig,
|
||||||
INITIAL_POSITION_TRIM_HORIZON, parentShardPollIntervalMillis, shardSyncIntervalMillis,
|
INITIAL_POSITION_TRIM_HORIZON, parentShardPollIntervalMillis, shardSyncIntervalMillis,
|
||||||
cleanupLeasesUponShardCompletion, leaseCoordinator, leaseCoordinator, executorService, metricsFactory,
|
cleanupLeasesUponShardCompletion, leaseCoordinator, leaseCoordinator, executorService, metricsFactory,
|
||||||
taskBackoffTimeMillis, failoverTimeMillis, false, shardPrioritization) {
|
taskBackoffTimeMillis, failoverTimeMillis, false, shardPrioritization) {
|
||||||
|
|
@ -888,10 +953,22 @@ public class WorkerTest {
|
||||||
|
|
||||||
when(coordinator.startGracefulShutdown(any(Callable.class))).thenReturn(gracefulShutdownFuture);
|
when(coordinator.startGracefulShutdown(any(Callable.class))).thenReturn(gracefulShutdownFuture);
|
||||||
|
|
||||||
Worker worker = new InjectableWorker("testRequestShutdown", recordProcessorFactory, streamConfig,
|
Worker worker = new InjectableWorker("testRequestShutdown",
|
||||||
INITIAL_POSITION_TRIM_HORIZON, parentShardPollIntervalMillis, shardSyncIntervalMillis,
|
recordProcessorFactory,
|
||||||
cleanupLeasesUponShardCompletion, leaseCoordinator, leaseCoordinator, executorService, metricsFactory,
|
config,
|
||||||
taskBackoffTimeMillis, failoverTimeMillis, false, shardPrioritization) {
|
streamConfig,
|
||||||
|
INITIAL_POSITION_TRIM_HORIZON,
|
||||||
|
parentShardPollIntervalMillis,
|
||||||
|
shardSyncIntervalMillis,
|
||||||
|
cleanupLeasesUponShardCompletion,
|
||||||
|
leaseCoordinator,
|
||||||
|
leaseCoordinator,
|
||||||
|
executorService,
|
||||||
|
metricsFactory,
|
||||||
|
taskBackoffTimeMillis,
|
||||||
|
failoverTimeMillis,
|
||||||
|
false,
|
||||||
|
shardPrioritization) {
|
||||||
@Override
|
@Override
|
||||||
void postConstruct() {
|
void postConstruct() {
|
||||||
this.gracefulShutdownCoordinator = coordinator;
|
this.gracefulShutdownCoordinator = coordinator;
|
||||||
|
|
@ -950,10 +1027,22 @@ public class WorkerTest {
|
||||||
when(recordProcessorFactory.createProcessor()).thenReturn(processor);
|
when(recordProcessorFactory.createProcessor()).thenReturn(processor);
|
||||||
|
|
||||||
|
|
||||||
Worker worker = new Worker("testRequestShutdown", recordProcessorFactory, streamConfig,
|
Worker worker = new Worker("testRequestShutdown",
|
||||||
INITIAL_POSITION_TRIM_HORIZON, parentShardPollIntervalMillis, shardSyncIntervalMillis,
|
recordProcessorFactory,
|
||||||
cleanupLeasesUponShardCompletion, leaseCoordinator, leaseCoordinator, executorService, metricsFactory,
|
config,
|
||||||
taskBackoffTimeMillis, failoverTimeMillis, false, shardPrioritization);
|
streamConfig,
|
||||||
|
INITIAL_POSITION_TRIM_HORIZON,
|
||||||
|
parentShardPollIntervalMillis,
|
||||||
|
shardSyncIntervalMillis,
|
||||||
|
cleanupLeasesUponShardCompletion,
|
||||||
|
leaseCoordinator,
|
||||||
|
leaseCoordinator,
|
||||||
|
executorService,
|
||||||
|
metricsFactory,
|
||||||
|
taskBackoffTimeMillis,
|
||||||
|
failoverTimeMillis,
|
||||||
|
false,
|
||||||
|
shardPrioritization);
|
||||||
|
|
||||||
when(executorService.submit(Matchers.<Callable<TaskResult>> any()))
|
when(executorService.submit(Matchers.<Callable<TaskResult>> any()))
|
||||||
.thenAnswer(new ShutdownHandlingAnswer(taskFuture));
|
.thenAnswer(new ShutdownHandlingAnswer(taskFuture));
|
||||||
|
|
@ -1020,10 +1109,22 @@ public class WorkerTest {
|
||||||
IRecordProcessor processor = mock(IRecordProcessor.class);
|
IRecordProcessor processor = mock(IRecordProcessor.class);
|
||||||
when(recordProcessorFactory.createProcessor()).thenReturn(processor);
|
when(recordProcessorFactory.createProcessor()).thenReturn(processor);
|
||||||
|
|
||||||
Worker worker = new Worker("testRequestShutdown", recordProcessorFactory, streamConfig,
|
Worker worker = new Worker("testRequestShutdown",
|
||||||
INITIAL_POSITION_TRIM_HORIZON, parentShardPollIntervalMillis, shardSyncIntervalMillis,
|
recordProcessorFactory,
|
||||||
cleanupLeasesUponShardCompletion, leaseCoordinator, leaseCoordinator, executorService, metricsFactory,
|
config,
|
||||||
taskBackoffTimeMillis, failoverTimeMillis, false, shardPrioritization);
|
streamConfig,
|
||||||
|
INITIAL_POSITION_TRIM_HORIZON,
|
||||||
|
parentShardPollIntervalMillis,
|
||||||
|
shardSyncIntervalMillis,
|
||||||
|
cleanupLeasesUponShardCompletion,
|
||||||
|
leaseCoordinator,
|
||||||
|
leaseCoordinator,
|
||||||
|
executorService,
|
||||||
|
metricsFactory,
|
||||||
|
taskBackoffTimeMillis,
|
||||||
|
failoverTimeMillis,
|
||||||
|
false,
|
||||||
|
shardPrioritization);
|
||||||
|
|
||||||
when(executorService.submit(Matchers.<Callable<TaskResult>> any()))
|
when(executorService.submit(Matchers.<Callable<TaskResult>> any()))
|
||||||
.thenAnswer(new ShutdownHandlingAnswer(taskFuture));
|
.thenAnswer(new ShutdownHandlingAnswer(taskFuture));
|
||||||
|
|
@ -1121,10 +1222,22 @@ public class WorkerTest {
|
||||||
IRecordProcessor processor = mock(IRecordProcessor.class);
|
IRecordProcessor processor = mock(IRecordProcessor.class);
|
||||||
when(recordProcessorFactory.createProcessor()).thenReturn(processor);
|
when(recordProcessorFactory.createProcessor()).thenReturn(processor);
|
||||||
|
|
||||||
Worker worker = new Worker("testRequestShutdown", recordProcessorFactory, streamConfig,
|
Worker worker = new Worker("testRequestShutdown",
|
||||||
INITIAL_POSITION_TRIM_HORIZON, parentShardPollIntervalMillis, shardSyncIntervalMillis,
|
recordProcessorFactory,
|
||||||
cleanupLeasesUponShardCompletion, leaseCoordinator, leaseCoordinator, executorService, metricsFactory,
|
config,
|
||||||
taskBackoffTimeMillis, failoverTimeMillis, false, shardPrioritization);
|
streamConfig,
|
||||||
|
INITIAL_POSITION_TRIM_HORIZON,
|
||||||
|
parentShardPollIntervalMillis,
|
||||||
|
shardSyncIntervalMillis,
|
||||||
|
cleanupLeasesUponShardCompletion,
|
||||||
|
leaseCoordinator,
|
||||||
|
leaseCoordinator,
|
||||||
|
executorService,
|
||||||
|
metricsFactory,
|
||||||
|
taskBackoffTimeMillis,
|
||||||
|
failoverTimeMillis,
|
||||||
|
false,
|
||||||
|
shardPrioritization);
|
||||||
|
|
||||||
when(executorService.submit(Matchers.<Callable<TaskResult>> any()))
|
when(executorService.submit(Matchers.<Callable<TaskResult>> any()))
|
||||||
.thenAnswer(new ShutdownHandlingAnswer(taskFuture));
|
.thenAnswer(new ShutdownHandlingAnswer(taskFuture));
|
||||||
|
|
@ -1226,10 +1339,22 @@ public class WorkerTest {
|
||||||
IRecordProcessor processor = mock(IRecordProcessor.class);
|
IRecordProcessor processor = mock(IRecordProcessor.class);
|
||||||
when(recordProcessorFactory.createProcessor()).thenReturn(processor);
|
when(recordProcessorFactory.createProcessor()).thenReturn(processor);
|
||||||
|
|
||||||
Worker worker = new Worker("testRequestShutdown", recordProcessorFactory, streamConfig,
|
Worker worker = new Worker("testRequestShutdown",
|
||||||
INITIAL_POSITION_TRIM_HORIZON, parentShardPollIntervalMillis, shardSyncIntervalMillis,
|
recordProcessorFactory,
|
||||||
cleanupLeasesUponShardCompletion, leaseCoordinator, leaseCoordinator, executorService, metricsFactory,
|
config,
|
||||||
taskBackoffTimeMillis, failoverTimeMillis, false, shardPrioritization);
|
streamConfig,
|
||||||
|
INITIAL_POSITION_TRIM_HORIZON,
|
||||||
|
parentShardPollIntervalMillis,
|
||||||
|
shardSyncIntervalMillis,
|
||||||
|
cleanupLeasesUponShardCompletion,
|
||||||
|
leaseCoordinator,
|
||||||
|
leaseCoordinator,
|
||||||
|
executorService,
|
||||||
|
metricsFactory,
|
||||||
|
taskBackoffTimeMillis,
|
||||||
|
failoverTimeMillis,
|
||||||
|
false,
|
||||||
|
shardPrioritization);
|
||||||
|
|
||||||
when(executorService.submit(Matchers.<Callable<TaskResult>> any()))
|
when(executorService.submit(Matchers.<Callable<TaskResult>> any()))
|
||||||
.thenAnswer(new ShutdownHandlingAnswer(taskFuture));
|
.thenAnswer(new ShutdownHandlingAnswer(taskFuture));
|
||||||
|
|
@ -1298,10 +1423,22 @@ public class WorkerTest {
|
||||||
IRecordProcessor processor = mock(IRecordProcessor.class);
|
IRecordProcessor processor = mock(IRecordProcessor.class);
|
||||||
when(recordProcessorFactory.createProcessor()).thenReturn(processor);
|
when(recordProcessorFactory.createProcessor()).thenReturn(processor);
|
||||||
|
|
||||||
Worker worker = new Worker("testRequestShutdown", recordProcessorFactory, streamConfig,
|
Worker worker = new Worker("testRequestShutdown",
|
||||||
INITIAL_POSITION_TRIM_HORIZON, parentShardPollIntervalMillis, shardSyncIntervalMillis,
|
recordProcessorFactory,
|
||||||
cleanupLeasesUponShardCompletion, leaseCoordinator, leaseCoordinator, executorService, metricsFactory,
|
config,
|
||||||
taskBackoffTimeMillis, failoverTimeMillis, false, shardPrioritization);
|
streamConfig,
|
||||||
|
INITIAL_POSITION_TRIM_HORIZON,
|
||||||
|
parentShardPollIntervalMillis,
|
||||||
|
shardSyncIntervalMillis,
|
||||||
|
cleanupLeasesUponShardCompletion,
|
||||||
|
leaseCoordinator,
|
||||||
|
leaseCoordinator,
|
||||||
|
executorService,
|
||||||
|
metricsFactory,
|
||||||
|
taskBackoffTimeMillis,
|
||||||
|
failoverTimeMillis,
|
||||||
|
false,
|
||||||
|
shardPrioritization);
|
||||||
|
|
||||||
when(executorService.submit(Matchers.<Callable<TaskResult>> any()))
|
when(executorService.submit(Matchers.<Callable<TaskResult>> any()))
|
||||||
.thenAnswer(new ShutdownHandlingAnswer(taskFuture));
|
.thenAnswer(new ShutdownHandlingAnswer(taskFuture));
|
||||||
|
|
@ -1336,16 +1473,29 @@ public class WorkerTest {
|
||||||
|
|
||||||
private abstract class InjectableWorker extends Worker {
|
private abstract class InjectableWorker extends Worker {
|
||||||
InjectableWorker(String applicationName, IRecordProcessorFactory recordProcessorFactory,
|
InjectableWorker(String applicationName, IRecordProcessorFactory recordProcessorFactory,
|
||||||
StreamConfig streamConfig, InitialPositionInStreamExtended initialPositionInStream,
|
KinesisClientLibConfiguration config, StreamConfig streamConfig,
|
||||||
|
InitialPositionInStreamExtended initialPositionInStream,
|
||||||
long parentShardPollIntervalMillis, long shardSyncIdleTimeMillis,
|
long parentShardPollIntervalMillis, long shardSyncIdleTimeMillis,
|
||||||
boolean cleanupLeasesUponShardCompletion, ICheckpoint checkpoint,
|
boolean cleanupLeasesUponShardCompletion, ICheckpoint checkpoint,
|
||||||
KinesisClientLibLeaseCoordinator leaseCoordinator, ExecutorService execService,
|
KinesisClientLibLeaseCoordinator leaseCoordinator, ExecutorService execService,
|
||||||
IMetricsFactory metricsFactory, long taskBackoffTimeMillis, long failoverTimeMillis,
|
IMetricsFactory metricsFactory, long taskBackoffTimeMillis, long failoverTimeMillis,
|
||||||
boolean skipShardSyncAtWorkerInitializationIfLeasesExist, ShardPrioritization shardPrioritization) {
|
boolean skipShardSyncAtWorkerInitializationIfLeasesExist, ShardPrioritization shardPrioritization) {
|
||||||
super(applicationName, recordProcessorFactory, streamConfig, initialPositionInStream,
|
super(applicationName,
|
||||||
parentShardPollIntervalMillis, shardSyncIdleTimeMillis, cleanupLeasesUponShardCompletion,
|
recordProcessorFactory,
|
||||||
checkpoint, leaseCoordinator, execService, metricsFactory, taskBackoffTimeMillis,
|
config,
|
||||||
failoverTimeMillis, skipShardSyncAtWorkerInitializationIfLeasesExist, shardPrioritization);
|
streamConfig,
|
||||||
|
initialPositionInStream,
|
||||||
|
parentShardPollIntervalMillis,
|
||||||
|
shardSyncIdleTimeMillis,
|
||||||
|
cleanupLeasesUponShardCompletion,
|
||||||
|
checkpoint,
|
||||||
|
leaseCoordinator,
|
||||||
|
execService,
|
||||||
|
metricsFactory,
|
||||||
|
taskBackoffTimeMillis,
|
||||||
|
failoverTimeMillis,
|
||||||
|
skipShardSyncAtWorkerInitializationIfLeasesExist,
|
||||||
|
shardPrioritization);
|
||||||
postConstruct();
|
postConstruct();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -1578,14 +1728,15 @@ public class WorkerTest {
|
||||||
lease.setCheckpoint(ExtendedSequenceNumber.AT_TIMESTAMP);
|
lease.setCheckpoint(ExtendedSequenceNumber.AT_TIMESTAMP);
|
||||||
initialLeases.add(lease);
|
initialLeases.add(lease);
|
||||||
}
|
}
|
||||||
runAndTestWorker(shardList, threadPoolSize, initialLeases, callProcessRecordsForEmptyRecordList, numberOfRecordsPerShard);
|
runAndTestWorker(shardList, threadPoolSize, initialLeases, callProcessRecordsForEmptyRecordList, numberOfRecordsPerShard, config);
|
||||||
}
|
}
|
||||||
|
|
||||||
private void runAndTestWorker(List<Shard> shardList,
|
private void runAndTestWorker(List<Shard> shardList,
|
||||||
int threadPoolSize,
|
int threadPoolSize,
|
||||||
List<KinesisClientLease> initialLeases,
|
List<KinesisClientLease> initialLeases,
|
||||||
boolean callProcessRecordsForEmptyRecordList,
|
boolean callProcessRecordsForEmptyRecordList,
|
||||||
int numberOfRecordsPerShard) throws Exception {
|
int numberOfRecordsPerShard,
|
||||||
|
KinesisClientLibConfiguration clientConfig) throws Exception {
|
||||||
File file = KinesisLocalFileDataCreator.generateTempDataFile(shardList, numberOfRecordsPerShard, "unitTestWT001");
|
File file = KinesisLocalFileDataCreator.generateTempDataFile(shardList, numberOfRecordsPerShard, "unitTestWT001");
|
||||||
IKinesisProxy fileBasedProxy = new KinesisLocalFileProxy(file.getAbsolutePath());
|
IKinesisProxy fileBasedProxy = new KinesisLocalFileProxy(file.getAbsolutePath());
|
||||||
|
|
||||||
|
|
@ -1597,7 +1748,7 @@ public class WorkerTest {
|
||||||
|
|
||||||
WorkerThread workerThread = runWorker(
|
WorkerThread workerThread = runWorker(
|
||||||
shardList, initialLeases, callProcessRecordsForEmptyRecordList, failoverTimeMillis,
|
shardList, initialLeases, callProcessRecordsForEmptyRecordList, failoverTimeMillis,
|
||||||
numberOfRecordsPerShard, fileBasedProxy, recordProcessorFactory, executorService, nullMetricsFactory);
|
numberOfRecordsPerShard, fileBasedProxy, recordProcessorFactory, executorService, nullMetricsFactory, clientConfig);
|
||||||
|
|
||||||
// TestStreamlet will release the semaphore once for every record it processes
|
// TestStreamlet will release the semaphore once for every record it processes
|
||||||
recordCounter.acquire(numberOfRecordsPerShard * shardList.size());
|
recordCounter.acquire(numberOfRecordsPerShard * shardList.size());
|
||||||
|
|
@ -1614,14 +1765,15 @@ public class WorkerTest {
|
||||||
}
|
}
|
||||||
|
|
||||||
private WorkerThread runWorker(List<Shard> shardList,
|
private WorkerThread runWorker(List<Shard> shardList,
|
||||||
List<KinesisClientLease> initialLeases,
|
List<KinesisClientLease> initialLeases,
|
||||||
boolean callProcessRecordsForEmptyRecordList,
|
boolean callProcessRecordsForEmptyRecordList,
|
||||||
long failoverTimeMillis,
|
long failoverTimeMillis,
|
||||||
int numberOfRecordsPerShard,
|
int numberOfRecordsPerShard,
|
||||||
IKinesisProxy kinesisProxy,
|
IKinesisProxy kinesisProxy,
|
||||||
IRecordProcessorFactory recordProcessorFactory,
|
IRecordProcessorFactory recordProcessorFactory,
|
||||||
ExecutorService executorService,
|
ExecutorService executorService,
|
||||||
IMetricsFactory metricsFactory) throws Exception {
|
IMetricsFactory metricsFactory,
|
||||||
|
KinesisClientLibConfiguration clientConfig) throws Exception {
|
||||||
final String stageName = "testStageName";
|
final String stageName = "testStageName";
|
||||||
final int maxRecords = 2;
|
final int maxRecords = 2;
|
||||||
|
|
||||||
|
|
@ -1653,6 +1805,7 @@ public class WorkerTest {
|
||||||
Worker worker =
|
Worker worker =
|
||||||
new Worker(stageName,
|
new Worker(stageName,
|
||||||
recordProcessorFactory,
|
recordProcessorFactory,
|
||||||
|
clientConfig,
|
||||||
streamConfig, INITIAL_POSITION_TRIM_HORIZON,
|
streamConfig, INITIAL_POSITION_TRIM_HORIZON,
|
||||||
parentShardPollIntervalMillis,
|
parentShardPollIntervalMillis,
|
||||||
shardSyncIntervalMillis,
|
shardSyncIntervalMillis,
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue