101 KiB
SpringBootBatch 源码
加载
版本使用 2.7.13
<dependency>
<groupId>org.springframework.boot</groupId>
<artifactId>spring-boot-dependencies</artifactId>
<version>2.7.13</version>
<scope>import</scope>
<type>pom</type>
</dependency>
spring-autoconfigure-metadata.properties
org.springframework.boot.autoconfigure.batch.BatchAutoConfiguration=
org.springframework.boot.autoconfigure.batch.BatchAutoConfiguration$DataSourceInitializerConfiguration=
org.springframework.boot.autoconfigure.batch.BatchAutoConfiguration$DataSourceInitializerConfiguration.ConditionalOnBean=javax.sql.DataSource
org.springframework.boot.autoconfigure.batch.BatchAutoConfiguration$DataSourceInitializerConfiguration.ConditionalOnClass=org.springframework.jdbc.datasource.init.DatabasePopulator
org.springframework.boot.autoconfigure.batch.BatchAutoConfiguration.AutoConfigureAfter=org.springframework.boot.autoconfigure.orm.jpa.HibernateJpaAutoConfiguration
org.springframework.boot.autoconfigure.batch.BatchAutoConfiguration.ConditionalOnBean=org.springframework.batch.core.launch.JobLauncher
org.springframework.boot.autoconfigure.batch.BatchAutoConfiguration.ConditionalOnClass=javax.sql.DataSource,org.springframework.batch.core.launch.JobLauncher
org.springframework.boot.autoconfigure.batch.BatchConfigurerConfiguration=
org.springframework.boot.autoconfigure.batch.BatchConfigurerConfiguration$JpaBatchConfiguration=
org.springframework.boot.autoconfigure.batch.BatchConfigurerConfiguration$JpaBatchConfiguration.ConditionalOnBean=
org.springframework.boot.autoconfigure.batch.BatchConfigurerConfiguration$JpaBatchConfiguration.ConditionalOnClass=javax.persistence.EntityManagerFactory
org.springframework.boot.autoconfigure.batch.BatchConfigurerConfiguration.ConditionalOnBean=javax.sql.DataSource
org.springframework.boot.autoconfigure.batch.BatchConfigurerConfiguration.ConditionalOnClass=org.springframework.transaction.PlatformTransactionManager
@EnableBatchProcessing
启动首先添加@EnableBatchProcessing
,这个类引入了BatchConfigurationSelector
.
BatchConfigurationSelector
这里面主要是判断modular
决定加载ModularBatchConfiguration
还是SimpleBatchConfiguration
SimpleBatchConfiguration
这个类继承了AbstractBatchConfiguration
。因为继承了InitializingBean
执行afterPropertiesSet
,创建出俩个类jobBuilderFactory
和 stepBuilderFactory
.
JobBuilderFactory
JobBuilderFactory 需要jobRepository
。SimpleBatchConfiguration#jobRepository()
方法,进入SimpleBatchConfiguration#createLazyProxy(AtomicReference<T> reference, Class<T> type)
。
是使用ProxyFactory
构建出代理对象,Advice
使用PassthruAdvice
类。
关键是TargetSource
属性,使用ReferenceTargetSource
类继承了AbstractLazyCreationTargetSource
,这是 aop 中TargetSource
,一种懒加载方式。在ReferenceTargetSource#createObject()
中调用了initialize
方法。
接着进入initialize
方法
/**
* Sets up the basic components by extracting them from the {@link BatchConfigurer configurer}, defaulting to some
* sensible values as long as a unique DataSource is available.
*
* @throws Exception if there is a problem in the configurer
*/
protected void initialize() throws Exception {
if (initialized) {
return;
}
BatchConfigurer configurer = getConfigurer(context.getBeansOfType(BatchConfigurer.class).values());
jobRepository.set(configurer.getJobRepository());
jobLauncher.set(configurer.getJobLauncher());
transactionManager.set(configurer.getTransactionManager());
jobRegistry.set(new MapJobRegistry());
jobExplorer.set(configurer.getJobExplorer());
initialized = true;
}
BatchConfigurer configurer=getConfigurer(context.getBeansOfType(BatchConfigurer.class).values());
protected BatchConfigurer getConfigurer(Collection<BatchConfigurer> configurers) throws Exception {
if (this.configurer != null) {
return this.configurer;
}
if (configurers == null || configurers.isEmpty()) {
if (dataSource == null) {
DefaultBatchConfigurer configurer = new DefaultBatchConfigurer();
configurer.initialize();
this.configurer = configurer;
return configurer;
} else {
DefaultBatchConfigurer configurer = new DefaultBatchConfigurer(dataSource);
configurer.initialize();
this.configurer = configurer;
return configurer;
}
}
if (configurers.size() > 1) {
throw new IllegalStateException(
"To use a custom BatchConfigurer the context must contain precisely one, found "
+ configurers.size());
}
this.configurer = configurers.iterator().next();
return this.configurer;
}
从 spring 中获取 BatchConfigurer
,没有使用 DefaultBatchConfigurer
,并调用 configurer.initialize()
- 获取所有的
BatchConfigurer
类,进入AbstractBatchConfiguration#getConfigurer
方法。如果为空就创建一个DefaultBatchConfigurer
,并执行configurer.initialize();
方法,DefaultBatchConfigurer
后面单独讲。 - 这里的
JobRepository
,JobLauncher
,JobRegistry
,PlatformTransactionManager
,JobExplorer
都是使用AtomicReference
原子性封装的。把initialized
设置为 true,只允许初始化一次
现在JobRepository
,JobLauncher
,JobRegistry
,PlatformTransactionManager
,JobExplorer
5 个类就全部初始化好了,
stepBuilderFactory
与上面一样,只不过初始化过了,就不会再初始化,直接获取值(注意:这里的都是使用AbstractLazyCreationTargetSource
创建代理对象,都是懒加载的)。
使用
首先弄清楚类直接关系,不使用数据库。所以所有默认,使用刚刚我们看过的源码默认创建的类就好了。
我们使用SimpleBatchConfiguration
进行讲解
这里需要设置spring.batch.job.enabled
为 false,不然会加载BatchAutoConfiguration#jobLauncherApplicationRunner
类,启动的时候会执行
job。BatchProperties
类放在BatchAutoConfiguration
类中讲解
BatchConfig
@Configuration
@EnableBatchProcessing
public class BatchConfig {
@Autowired
JobBuilderFactory jobBuilders;
@Autowired
StepBuilderFactory stepBuilders;
@Bean
public Job footballjob() {
return jobBuilders.get("test-job")
.start(step())
.build();
}
@Bean
public Step step() {
return stepBuilders.get("test-step")
.<Integer, Integer>chunk(2)
.reader(itemReader())
.writer(itemWriter())
.build();
}
@Bean
public ItemReader<Integer> itemReader() {
return new ItemReader<Integer>() {
final AtomicInteger atomicInteger = new AtomicInteger();
@Override
public Integer read() throws UnexpectedInputException, ParseException, NonTransientResourceException {
if (atomicInteger.compareAndSet(10, 0)) {
return null;
}
atomicInteger.getAndIncrement();
return atomicInteger.intValue();
}
};
}
@Bean
public ItemWriter<Integer> itemWriter() {
return new ItemWriter<Integer>() {
@Override
public void write(List<? extends Integer> items) {
System.out.println("一次读取:" + Arrays.toString(items.toArray()));
}
};
}
}
Step
使用JobBuilderFactory
创建出Step
。JobBuilderFactory
在上面AbstractBatchConfiguration
中创建了。
@Bean
public Step step() {
TaskletStep taskletStep = stepBuilders.get("test-step")
.<Integer, Integer>chunk(2)
.reader(itemReader())
.writer(itemWriter())
.listener(new StepExecutionListener() {
@Override
public void beforeStep(StepExecution stepExecution) {
System.out.println("step-我是beforeJob--" + stepExecution);
}
@Override
public ExitStatus afterStep(StepExecution stepExecution) {
System.out.println("step-我是beforeJob--" + stepExecution);
return stepExecution.getExitStatus();
}
})
// .startLimit(2)
.build();
taskletStep.setAllowStartIfComplete(true);
return taskletStep;
}
stepBuilders.get("test-stop")
StepBuilderFactory
只有一个方法 get
public StepBuilder get(String name) {
StepBuilder builder = new StepBuilder(name).repository(jobRepository).transactionManager(
transactionManager);
return builder;
}
stepBuilders.get("test-stop")
创建一个StepBuilder
。
StepBuilder
继承StepBuilderHelper<StepBuilder>
,StepBuilderHelper
中有个protected final CommonStepProperties properties;
属性。
对CommonStepProperties
中name
,JobRepository
和PlatformTransactionManager
进行赋值。
<Integer, Integer>chunk(2)
/**
* Build a step that processes items in chunks with the size provided. To extend the step to being fault tolerant,
* call the {@link SimpleStepBuilder#faultTolerant()} method on the builder. In most cases you will want to
* parameterize your call to this method, to preserve the type safety of your readers and writers, e.g.
*
* <pre>
* new StepBuilder("step1").<Order, Ledger> chunk(100).reader(new OrderReader()).writer(new LedgerWriter())
* // ... etc.
* </pre>
*
* @param chunkSize the chunk size (commit interval)
* @return a {@link SimpleStepBuilder}
* @param <I> the type of item to be processed as input
* @param <O> the type of item to be output
*/
public<I, O> SimpleStepBuilder<I, O> chunk(int chunkSize){
return new SimpleStepBuilder<I, O>(this).chunk(chunkSize);
}
创建一个SimpleStepBuilder
,并确定泛型,设置一个chunkSize
大小。翻译一下注释构建一个步骤,以提供的大小分块处理项目。要将步骤扩展为容错,请在构建器上调用SimpleStepBuilder.faultTolerant()方法。在大多数情况下,您需要参数化对此方法的调用,以保护读者和作者的类型安全
,chunkSize – 块大小(提交间隔)。
接下来都是操作SimpleStepBuilder
类。这个大小也就是是每次读取多少次。
SimpleStepBuilder
继承AbstractTaskletStepBuilder<SimpleStepBuilder<I, O>>
类。AbstractTaskletStepBuilder<SimpleStepBuilder<I, O>>
也是继承StepBuilderHelper<AbstractTaskletStepBuilder<B>>
。
.reader(itemReader())
这里就是读取,需要一个ItemReader<? extends I> reader
对象。
创建一个
@Bean
public ItemReader<Integer> itemReader() {
return new ItemReader<Integer>() {
final AtomicInteger atomicInteger = new AtomicInteger();
@Override
public Integer read() throws UnexpectedInputException, ParseException, NonTransientResourceException {
if (atomicInteger.compareAndSet(10, 0)) {
return null;
}
atomicInteger.getAndIncrement();
return atomicInteger.intValue();
}
};
}
每次读取 1 个值,总共读取 10 个值。返回 null 表示没有数据了。
.writer(itemWriter())
与上面一样。这里的泛型要注意,是根据.<Integer, Integer>chunk(2)
进行确认的,前面 Integer 表示reader
返回值,后面表示writer
入参。
@Bean
public ItemWriter<Integer> itemWriter() {
return new ItemWriter<Integer>() {
@Override
public void write(List<? extends Integer> items) {
System.out.println("一次读取:" + Arrays.toString(items.toArray()));
}
};
}
.build();
/**
* Build a step with the reader, writer, processor as provided.
*
* @see org.springframework.batch.core.step.builder.AbstractTaskletStepBuilder#build()
*/
@Override
public TaskletStep build() {
registerStepListenerAsItemListener();
registerAsStreamsAndListeners(reader, processor, writer);
return super.build();
}
registerStepListenerAsItemListener()
:注册监听器,ItemReadListener
,ItemProcessListener
,ItemWriteListener
.
registerAsStreamsAndListeners
:注册ItemStream
,StepListenerMetaData
中的所有接口,StepExecutionListener
,ChunkListener
,ItemReadListener
,ItemProcessListener
和ItemWriteListener
.
AbstractListenerFactoryBean.isListener(delegate, StepListener.class, StepListenerMetaData.values());
进行判断
BEFORE_STEP("beforeStep","before-step-method",BeforeStep.class,StepExecutionListener.class,StepExecution.class),
AFTER_STEP("afterStep","after-step-method",AfterStep.class,StepExecutionListener.class,StepExecution.class),
BEFORE_CHUNK("beforeChunk","before-chunk-method",BeforeChunk.class,ChunkListener.class,ChunkContext.class),
AFTER_CHUNK("afterChunk","after-chunk-method",AfterChunk.class,ChunkListener.class,ChunkContext.class),
AFTER_CHUNK_ERROR("afterChunkError","after-chunk-error-method",AfterChunkError.class,ChunkListener.class,ChunkContext.class),
BEFORE_READ("beforeRead","before-read-method",BeforeRead.class,ItemReadListener.class),
AFTER_READ("afterRead","after-read-method",AfterRead.class,ItemReadListener.class,Object.class),
ON_READ_ERROR("onReadError","on-read-error-method",OnReadError.class,ItemReadListener.class,Exception.class),
BEFORE_PROCESS("beforeProcess","before-process-method",BeforeProcess.class,ItemProcessListener.class,Object.class),
AFTER_PROCESS("afterProcess","after-process-method",AfterProcess.class,ItemProcessListener.class,Object.class,Object.class),
ON_PROCESS_ERROR("onProcessError","on-process-error-method",OnProcessError.class,ItemProcessListener.class,Object.class,Exception.class),
BEFORE_WRITE("beforeWrite","before-write-method",BeforeWrite.class,ItemWriteListener.class,List.class),
AFTER_WRITE("afterWrite","after-write-method",AfterWrite.class,ItemWriteListener.class,List.class),
ON_WRITE_ERROR("onWriteError","on-write-error-method",OnWriteError.class,ItemWriteListener.class,Exception.class,List.class),
ON_SKIP_IN_READ("onSkipInRead","on-skip-in-read-method",OnSkipInRead.class,SkipListener.class,Throwable.class),
ON_SKIP_IN_PROCESS("onSkipInProcess","on-skip-in-process-method",OnSkipInProcess.class,SkipListener.class,Object.class,Throwable.class),
ON_SKIP_IN_WRITE("onSkipInWrite","on-skip-in-write-method",OnSkipInWrite.class,SkipListener.class,Object.class,Throwable.class);
这些监听器之类的后面单独讲,所有这些之类的顶级接口StepListener
。
super.build();
public TaskletStep build() {
registerStepListenerAsChunkListener();
TaskletStep step = new TaskletStep(getName());
super.enhance(step);
step.setChunkListeners(chunkListeners.toArray(new ChunkListener[0]));
if (transactionAttribute != null) {
step.setTransactionAttribute(transactionAttribute);
}
if (stepOperations == null) {
stepOperations = new RepeatTemplate();
if (taskExecutor != null) {
TaskExecutorRepeatTemplate repeatTemplate = new TaskExecutorRepeatTemplate();
repeatTemplate.setTaskExecutor(taskExecutor);
repeatTemplate.setThrottleLimit(throttleLimit);
stepOperations = repeatTemplate;
}
((RepeatTemplate) stepOperations).setExceptionHandler(exceptionHandler);
}
step.setStepOperations(stepOperations);
step.setTasklet(createTasklet());
step.setStreams(streams.toArray(new ItemStream[0]));
try {
step.afterPropertiesSet();
} catch (Exception e) {
throw new StepBuilderException(e);
}
return step;
}
super.enhance(step);
把CommonStepProperties
中的jobRepository
,allowStartIfComplete
,startLimit
和stepExecutionListeners
赋值给step
.
RepeatOperations
注释翻译后:批处理模板实现的简单实现 RepeatOperations 和基类。提供包含拦截器和策略的框架。子类只需要提供一个获取下一个结果的方法,以及一个等待所有结果从并发进程或线程返回的方法。 注意:模板在迭代过程中累积抛出的异常,当主循环结束时(即完成对项目的处理),它们都会一起处理。不希望在引发异常时停止执行的客户端可以使用在收到异常时未完成的特定 CompletionPolicy 。这不是默认行为。 想要在 引发 RepeatCallback 异常时执行某些业务操作的客户端可以考虑使用自定义,而不是尝试自定义 RepeatListener CompletionPolicy.这通常是一个更友好的接口来实现,该方法在回调的结果中传递,如果业务处理抛出异常, RepeatListener.after( RepeatContext, RepeatStatus) 这将是一个实例 Throwable 。 如果不将异常传播到调用方,则还需要提供非默认值 CompletionPolicy ,但这可能是现成的,业务操作仅在拦截器中实现。
taskExecutor:一般用来设置线程池线程池。这里我们也没有设置这个值。
((RepeatTemplate) stepOperations).setExceptionHandler(exceptionHandler);:异常处理,默认抛出异常,DefaultExceptionHandler
类。
step.setStepOperations(stepOperations);
赋值给 step
step.setTasklet(createTasklet());
@Override
protected Tasklet createTasklet() {
Assert.state(reader != null, "ItemReader must be provided");
Assert.state(writer != null, "ItemWriter must be provided");
RepeatOperations repeatOperations = createChunkOperations();
SimpleChunkProvider<I> chunkProvider = new SimpleChunkProvider<>(getReader(), repeatOperations);
SimpleChunkProcessor<I, O> chunkProcessor = new SimpleChunkProcessor<>(getProcessor(), getWriter());
chunkProvider.setListeners(new ArrayList<>(itemListeners));
chunkProcessor.setListeners(new ArrayList<>(itemListeners));
ChunkOrientedTasklet<I> tasklet = new ChunkOrientedTasklet<>(chunkProvider, chunkProcessor);
tasklet.setBuffering(!readerTransactionalQueue);
return tasklet;
}
createChunkOperations();
:重复处理类,这里也是RepeatTemplate
。
protected RepeatOperations createChunkOperations() {
RepeatOperations repeatOperations = chunkOperations;
if (repeatOperations == null) {
RepeatTemplate repeatTemplate = new RepeatTemplate();
repeatTemplate.setCompletionPolicy(getChunkCompletionPolicy());
repeatOperations = repeatTemplate;
}
return repeatOperations;
}
里面的repeatTemplate.setCompletionPolicy(getChunkCompletionPolicy());
。
/**
* @return a {@link CompletionPolicy} consistent with the chunk size and injected policy (if present).
*/
protected CompletionPolicy getChunkCompletionPolicy() {
Assert.state(!(completionPolicy != null && chunkSize > 0),
"You must specify either a chunkCompletionPolicy or a commitInterval but not both.");
Assert.state(chunkSize >= 0, "The commitInterval must be positive or zero (for default value).");
if (completionPolicy != null) {
return completionPolicy;
}
if (chunkSize == 0) {
if (logger.isInfoEnabled()) {
logger.info("Setting commit interval to default value (" + DEFAULT_COMMIT_INTERVAL + ")");
}
chunkSize = DEFAULT_COMMIT_INTERVAL;
}
return new SimpleCompletionPolicy(chunkSize);
}
每次处理大小:在固定数量的操作后终止批处理的策略。维护内部状态并增加计数器,因此成功使用此策略需要每个批处理项仅调用一次isComplete()。使用标准的RepeatTemplate应确保保留此合同,但需要仔细监控。
留心观察一下就会发现,这个值是.<Integer, Integer>chunk(2)
这里的
2,里面还可以传一个 chunk(CompletionPolicy completionPolicy)
。
SimpleStepBuilder#createTasklet()
创建一个ChunkOrientedTasklet
。里面有俩个属性chunkProvider
和chunkProcessor
。
SimpleChunkProvider chunkProvider = new SimpleChunkProvider<>(getReader(), repeatOperations);
然后创建SimpleChunkProvider
对象,getReader()
就是ItemReader
,repeatOperations
就是SimpleCompletionPolicy
。
impleChunkProcessor<I, O> chunkProcessor = new SimpleChunkProcessor<>(getProcessor(), getWriter());
getProcessor()
:就是processor(Function<? super I, ? extends O> function)
,stepBuilders.get().processor()
里面的值。
使用chunkProvider, chunkProcessor
构建出ChunkOrientedTasklet
返回给step
的tasklet
。
/**
* Public setter for the {@link Tasklet}.
*
* @param tasklet the {@link Tasklet} to set
*/
public void setTasklet(Tasklet tasklet) {
this.tasklet = tasklet;
if (tasklet instanceof StepExecutionListener) {
registerStepExecutionListener((StepExecutionListener) tasklet);
}
}
这里默认构建出来的不是StepExecutionListener
。是ChunkOrientedTasklet
。这里不是主体流程,下面补充。
step.afterPropertiesSet();
里面就是一些校验了。
Job
使用JobBuilderFactory
创建出Step
。JobBuilderFactory
在上面AbstractBatchConfiguration
创建了。
@Bean
public Job footballjob() {
return jobBuilders.get("test")
.start(step())
.build();
}
jobBuilders.get("test")
与 step 基本一样,先创建JobBuilder
。 Flow
下面单独讲。
.start(step())
/**
* Create a new job builder that will execute a step or sequence of steps.
*
* @param step a step to execute
* @return a {@link SimpleJobBuilder}
*/
public SimpleJobBuilder start(Step step) {
return new SimpleJobBuilder(this).start(step);
}
创建一个SimpleJobBuilder
,set 属性 private List<Step> steps = new ArrayList<>();
。
.build();
public Job build() {
if (builder != null) {
return builder.end().build();
}
SimpleJob job = new SimpleJob(getName());
super.enhance(job);
job.setSteps(steps);
try {
job.afterPropertiesSet();
} catch (Exception e) {
throw new JobBuilderException(e);
}
return job;
}
真正构建出来的是SimpleJob
对象。
super.enhance(job);
protected void enhance(Job target) {
if (target instanceof AbstractJob) {
AbstractJob job = (AbstractJob) target;
job.setJobRepository(properties.getJobRepository());
JobParametersIncrementer jobParametersIncrementer = properties.getJobParametersIncrementer();
if (jobParametersIncrementer != null) {
job.setJobParametersIncrementer(jobParametersIncrementer);
}
JobParametersValidator jobParametersValidator = properties.getJobParametersValidator();
if (jobParametersValidator != null) {
job.setJobParametersValidator(jobParametersValidator);
}
Boolean restartable = properties.getRestartable();
if (restartable != null) {
job.setRestartable(restartable);
}
List<JobExecutionListener> listeners = properties.getJobExecutionListeners();
if (!listeners.isEmpty()) {
job.setJobExecutionListeners(listeners.toArray(new JobExecutionListener[0]));
}
}
}
把CommonJobProperties properties
的属性赋值,比如JobRepository
,JobParametersIncrementer
,JobParametersValidator
和Restartable
.
job.setJobExecutionListeners(listeners.toArray(new JobExecutionListener[0]));
List<JobExecutionListener> listeners = properties.getJobExecutionListeners();
if (!listeners.isEmpty()) {
job.setJobExecutionListeners(listeners.toArray(new JobExecutionListener[0]));
}
获取里面的JobExecutionListener
set。
job.setSteps(steps);
里面所有的steps
。
job.afterPropertiesSet();
一些校验
JobLauncher
真正的使用类,在 spring-boot 中ApplicationRunner
也是使用这个JobLauncherApplicationRunner#executeRegisteredJobs(JobParameters jobParameters)
。
@RestController
public class BatchController {
@Autowired
JobLauncher jobLauncher;
@Autowired
Job footballjob;
@GetMapping("/batch")
public String getBatch() throws JobInstanceAlreadyCompleteException, JobExecutionAlreadyRunningException, JobParametersInvalidException, JobRestartException {
JobParameters jobParameters = new JobParametersBuilder().toJobParameters();
jobLauncher.run(footballjob, jobParameters);
return "batch";
}
}
JobParameters
下面单独讲解。
JobLauncher
对象在SimpleBatchConfiguration
中使用懒加载的代理对象创建出来的。
Job
是上面我们自己创建的。
进入到SimpleJobLauncher#run(final Job job, final JobParameters jobParameters)
中。
JobExecution lastExecution = jobRepository.getLastJobExecution(job.getName(), jobParameters);
@Override
@Nullable
public JobExecution getLastJobExecution(String jobName, JobParameters jobParameters) {
JobInstance jobInstance = jobInstanceDao.getJobInstance(jobName, jobParameters);
if (jobInstance == null) {
return null;
}
JobExecution jobExecution = jobExecutionDao.getLastJobExecution(jobInstance);
if (jobExecution != null) {
jobExecution.setExecutionContext(ecDao.getExecutionContext(jobExecution));
stepExecutionDao.addStepExecutions(jobExecution);
}
return jobExecution;
}
MapJobInstanceDao
已弃用 从 v4.3 开始,赞成使用 与 JdbcJobInstanceDao 内存数据库一起使用。计划在 v5.0 中删除。
MapJobInstanceDao#getJobInstance
,第一次进来是获取不到的,直接返回 null。
job.getJobParametersValidator().validate(jobParameters);
校验参数。我们没填。
jobExecution = jobRepository.createJobExecution(job.getName(), jobParameters);
创建一个JobExecution
SimpleJobRepository#createJobExecution(String jobName, JobParameters jobParameters)
@Override
public JobExecution createJobExecution(String jobName, JobParameters jobParameters) {
JobInstance jobInstance = jobInstanceDao.getJobInstance(jobName, jobParameters);
ExecutionContext executionContext;
if (jobInstance != null) {
List<JobExecution> executions = jobExecutionDao.findJobExecutions(jobInstance);
if (executions.isEmpty()) {
throw new IllegalStateException("Cannot find any job execution for job instance: " + jobInstance);
}
for (JobExecution execution : executions) {
if (execution.isRunning() || execution.isStopping()) {
throw new JobExecutionAlreadyRunningException("A job execution for this job is already running: "
+ jobInstance);
}
BatchStatus status = execution.getStatus();
if (status == BatchStatus.UNKNOWN) {
throw new JobRestartException("Cannot restart job from UNKNOWN status. "
+ "The last execution ended with a failure that could not be rolled back, "
+ "so it may be dangerous to proceed. Manual intervention is probably necessary.");
}
Collection<JobParameter> allJobParameters = execution.getJobParameters().getParameters().values();
long identifyingJobParametersCount = allJobParameters.stream().filter(JobParameter::isIdentifying).count();
if (identifyingJobParametersCount > 0 && (status == BatchStatus.COMPLETED || status == BatchStatus.ABANDONED)) {
throw new JobInstanceAlreadyCompleteException(
"A job instance already exists and is complete for parameters=" + jobParameters
+ ". If you want to run this job again, change the parameters.");
}
}
executionContext = ecDao.getExecutionContext(jobExecutionDao.getLastJobExecution(jobInstance));
} else {
// no job found, create one
jobInstance = jobInstanceDao.createJobInstance(jobName, jobParameters);
executionContext = new ExecutionContext();
}
JobExecution jobExecution = new JobExecution(jobInstance, jobParameters, null);
jobExecution.setExecutionContext(executionContext);
jobExecution.setLastUpdated(new Date(System.currentTimeMillis()));
// Save the JobExecution so that it picks up an ID (useful for clients
// monitoring asynchronous executions):
jobExecutionDao.saveJobExecution(jobExecution);
ecDao.saveExecutionContext(jobExecution);
return jobExecution;
}
首先JobInstance jobInstance = jobInstanceDao.getJobInstance(jobName, jobParameters);
。这里的jobInstanceDao
是
Mapxxx,注意这几个类都是过时的,我们先把流程弄懂在看这些细节。
为空,去到else
。
jobInstanceDao#createJobInstance(String jobName, JobParameters jobParameters)
@Override
public JobInstance createJobInstance(String jobName, JobParameters jobParameters) {
Assert.state(getJobInstance(jobName, jobParameters) == null, "JobInstance must not already exist");
JobInstance jobInstance = new JobInstance(currentId.getAndIncrement(), jobName);
jobInstance.incrementVersion();
jobInstances.put(jobName + "|" + jobKeyGenerator.generateKey(jobParameters), jobInstance);
return jobInstance;
}
就是Map<String, JobInstance> jobInstances = new ConcurrentHashMap<>();
put
了一个值。后面jobKeyGenerator.generateKey(jobParameters)
默认为一个 md5 加密算法。
现在MapJobInstanceDao#jobInstances
就有值了。接着回到SimpleJobRepository
类。
这里有个地方需要注意,这些类都是一些代理对象,使用代理类懒加载的。断点打在实现类上就好了!
executionContext = new ExecutionContext();
里面维护了一个dirty
和Map<String, Object> map
。
JobExecution
接着创建一个JobExecution
对象。继承了Entity
对象。
把ExecutionContext
赋值。
最后修改时间赋值为当前时间。
这个类里面关注jobParameters
和jobInstance
对象。jobInstanceDao
创建的jobInstance
对象。
jobExecutionDao.saveJobExecution(jobExecution);
Assert.isTrue(jobExecution.getId() == null, "jobExecution id is not null");
Long newId = currentId.getAndIncrement();
jobExecution.setId(newId);
jobExecution.incrementVersion();
executionsById.put(newId, copy(jobExecution));
来到MapJobExecutionDao
类,这 4 个Dao
后面单独讲解。现在都是使用 MapxxxDao。
里面有一个ConcurrentMap<Long, JobExecution> executionsById = new ConcurrentHashMap<>();
和AtomicLong currentId = new AtomicLong(0L);
版本号和值加一,并且存入 map 中,如果是数据库,就会存入数据库。逻辑基本一致。
ecDao.saveExecutionContext(jobExecution);
@Override
public void saveExecutionContext(JobExecution jobExecution) {
updateExecutionContext(jobExecution);
}
@Override
public void updateExecutionContext(JobExecution jobExecution) {
ExecutionContext executionContext = jobExecution.getExecutionContext();
if (executionContext != null) {
contexts.put(ContextKey.job(jobExecution.getId()), copy(executionContext));
}
}
值放入ConcurrentMap<ContextKey, ExecutionContext> contexts = TransactionAwareProxyFactory.createAppendOnlyTransactionalMap()
中
taskExecutor.execute(new Runnable() )
try {
taskExecutor.execute(new Runnable() {
@Override
public void run() {
try {
job.execute(jobExecution);
} catch (Throwable t) {
rethrow(t);
}
}
private void rethrow(Throwable t) {
if (t instanceof RuntimeException) {
throw (RuntimeException) t;
} else if (t instanceof Error) {
throw (Error) t;
}
throw new IllegalStateException(t);
}
});
} catch (TaskRejectedException e) {
jobExecution.upgradeStatus(BatchStatus.FAILED);
if (jobExecution.getExitStatus().equals(ExitStatus.UNKNOWN)) {
jobExecution.setExitStatus(ExitStatus.FAILED.addExitDescription(e));
}
jobRepository.update(jobExecution);
}
taskExecutor
是在afterPropertiesSet()
方法创建的,为SyncTaskExecutor
。这个是在DefaultBatchConfigurer#createJobLauncher()
中调用。
这里主要就是俩行代码
执行job.execute(jobExecution);
,出现异常执行rethrow(t);
。
job.execute(jobExecution execution)
job 是SimpleJob
类。进入AbstractJob#execute(JobExecution execution)
方法。
@Override
public final void execute(JobExecution execution) {
Assert.notNull(execution, "jobExecution must not be null");
if (logger.isDebugEnabled()) {
logger.debug("Job execution starting: " + execution);
}
JobSynchronizationManager.register(execution);
LongTaskTimer longTaskTimer = BatchMetrics.createLongTaskTimer("job.active", "Active jobs",
Tag.of("name", execution.getJobInstance().getJobName()));
LongTaskTimer.Sample longTaskTimerSample = longTaskTimer.start();
Timer.Sample timerSample = BatchMetrics.createTimerSample();
try {
jobParametersValidator.validate(execution.getJobParameters());
if (execution.getStatus() != BatchStatus.STOPPING) {
execution.setStartTime(new Date());
updateStatus(execution, BatchStatus.STARTED);
listener.beforeJob(execution);
try {
doExecute(execution);
if (logger.isDebugEnabled()) {
logger.debug("Job execution complete: " + execution);
}
} catch (RepeatException e) {
throw e.getCause();
}
} else {
// The job was already stopped before we even got this far. Deal
// with it in the same way as any other interruption.
execution.setStatus(BatchStatus.STOPPED);
execution.setExitStatus(ExitStatus.COMPLETED);
if (logger.isDebugEnabled()) {
logger.debug("Job execution was stopped: " + execution);
}
}
} catch (JobInterruptedException e) {
if (logger.isInfoEnabled()) {
logger.info("Encountered interruption executing job: "
+ e.getMessage());
}
if (logger.isDebugEnabled()) {
logger.debug("Full exception", e);
}
execution.setExitStatus(getDefaultExitStatusForFailure(e, execution));
execution.setStatus(BatchStatus.max(BatchStatus.STOPPED, e.getStatus()));
execution.addFailureException(e);
} catch (Throwable t) {
logger.error("Encountered fatal error executing job", t);
execution.setExitStatus(getDefaultExitStatusForFailure(t, execution));
execution.setStatus(BatchStatus.FAILED);
execution.addFailureException(t);
} finally {
try {
if (execution.getStatus().isLessThanOrEqualTo(BatchStatus.STOPPED)
&& execution.getStepExecutions().isEmpty()) {
ExitStatus exitStatus = execution.getExitStatus();
ExitStatus newExitStatus =
ExitStatus.NOOP.addExitDescription("All steps already completed or no steps configured for this job.");
execution.setExitStatus(exitStatus.and(newExitStatus));
}
timerSample.stop(BatchMetrics.createTimer("job", "Job duration",
Tag.of("name", execution.getJobInstance().getJobName()),
Tag.of("status", execution.getExitStatus().getExitCode())
));
longTaskTimerSample.stop();
execution.setEndTime(new Date());
try {
listener.afterJob(execution);
} catch (Exception e) {
logger.error("Exception encountered in afterJob callback", e);
}
jobRepository.update(execution);
} finally {
JobSynchronizationManager.release();
}
}
}
JobSynchronizationManager.register(execution);
@Nullable
public C register(@Nullable E execution) {
if (execution == null) {
return null;
}
getCurrent().push(execution);
C context;
synchronized (contexts) {
context = contexts.get(execution);
if (context == null) {
context = createNewContext(execution, null);
contexts.put(execution, context);
}
}
increment();
return context;
}
SynchronizationManagerSupport
的泛型为JobExecution, JobContext
/**
* 当前执行的存储; 必须是ThreadLocal的,因为它需要在不属于步骤/作业的组件中定位上下文 (如重新补水作用域代理时)。不使用InheritableThreadLocal,因为如果一个步骤试图运行多个子步骤 (例如分区),会有副作用。
* 堆栈用于覆盖单线程的情况,从而使API与多线程相同。
*/
private final ThreadLocal<Stack<E>>executionHolder=new ThreadLocal<>();
/**
* 每个执行的引用计数器: 有多少个线程正在使用同一个?
*/
private final Map<E, AtomicInteger> counts=new ConcurrentHashMap<>();
/**
* 从正在运行的执行到关联上下文的简单映射。
*/
private final Map<E, C> contexts=new ConcurrentHashMap<>();
increment();
public void increment() {
E current = getCurrent().peek();
if (current != null) {
AtomicInteger count;
synchronized (counts) {
count = counts.get(current);
if (count == null) {
count = new AtomicInteger();
counts.put(current, count);
}
}
count.incrementAndGet();
}
}
简单来说就是里面三个属性不存在就创建。存在把executionHolder
更新。count
++。
SynchronizationManagerSupport
类维护了三个对象
/**
* Storage for the current execution; has to be ThreadLocal because it
* is needed to locate a context in components that are not part of a
* step/job (like when re-hydrating a scoped proxy). Doesn't use
* InheritableThreadLocal because there are side effects if a step is trying
* to run multiple child steps (e.g. with partitioning). The Stack is used
* to cover the single threaded case, so that the API is the same as
* multi-threaded.
*/
private final ThreadLocal<Stack<E>>executionHolder=new ThreadLocal<>();
/**
* Reference counter for each execution: how many threads are using the
* same one?
*/
private final Map<E, AtomicInteger> counts=new ConcurrentHashMap<>();
/**
* Simple map from a running execution to the associated context.
*/
private final Map<E, C> contexts=new ConcurrentHashMap<>();
接着回到AbstractJob
类
LongTaskTimer
:计时器。
jobParametersValidator.validate(execution.getJobParameters());
参数没有,跳过
if (execution.getStatus() != BatchStatus.STOPPING) {
默认状态为BatchStatus.STARTING
。
execution.setStartTime(new Date());
设置当前开始时间
updateStatus(execution, BatchStatus.STARTED);
private void updateStatus(JobExecution jobExecution, BatchStatus status) {
jobExecution.setStatus(status);
jobRepository.update(jobExecution);
}
jobExecution.setStatus(status);
设置当前状态.
jobRepository.update(jobExecution);
更新当前jobExecution
。
SimpleJobRepository#update(JobExecution jobExecution)
@Override
public void update(JobExecution jobExecution) {
Assert.notNull(jobExecution, "JobExecution cannot be null.");
Assert.notNull(jobExecution.getJobId(), "JobExecution must have a Job ID set.");
Assert.notNull(jobExecution.getId(), "JobExecution must be already saved (have an id assigned).");
jobExecution.setLastUpdated(new Date(System.currentTimeMillis()));
jobExecutionDao.synchronizeStatus(jobExecution);
if (jobExecution.getStatus() == BatchStatus.STOPPING && jobExecution.getEndTime() != null) {
if (logger.isInfoEnabled()) {
logger.info("Upgrading job execution status from STOPPING to STOPPED since it has already ended.");
}
jobExecution.upgradeStatus(BatchStatus.STOPPED);
}
jobExecutionDao.updateJobExecution(jobExecution);
}
先判断不能为空,设置最后修改时间。
jobExecutionDao.synchronizeStatus(jobExecution);
MapJobExecutionDao#synchronizeStatus(JobExecution jobExecution)
@Override
public void synchronizeStatus(JobExecution jobExecution) {
JobExecution saved = getJobExecution(jobExecution.getId());
if (saved.getVersion().intValue() != jobExecution.getVersion().intValue()) {
jobExecution.upgradeStatus(saved.getStatus());
jobExecution.setVersion(saved.getVersion());
}
}
版本不一样就更新,然后版本++,这里版本没有改变。回到SimpleJobRepository#update(JobExecution jobExecution)
。如果状态STOPPING
并且结束时间不为空,就使用jobExecution
更新。这里也没有结束,不需要更新。
jobExecutionDao.updateJobExecution(jobExecution);
MapJobExecutionDao#updateJobExecution(JobExecution jobExecution)
@Override
public void updateJobExecution(JobExecution jobExecution) {
Long id = jobExecution.getId();
Assert.notNull(id, "JobExecution is expected to have an id (should be saved already)");
JobExecution persistedExecution = executionsById.get(id);
Assert.notNull(persistedExecution, "JobExecution must already be saved");
synchronized (jobExecution) {
if (!persistedExecution.getVersion().equals(jobExecution.getVersion())) {
throw new OptimisticLockingFailureException("Attempt to update job execution id=" + id
+ " with wrong version (" + jobExecution.getVersion() + "), where current version is "
+ persistedExecution.getVersion());
}
jobExecution.incrementVersion();
executionsById.put(id, copy(jobExecution));
}
}
首先校验不能为空,版本不一致就抛出异常,这里基本都是使用乐观锁。官网有介绍。接着更新版本,保存数据。
到此这一步结束了,主要是跟新版本,把版本从STARTING
跟新为BatchStatus.STARTED
。
回到AbstractJob#execute(JobExecution execution)
,接着执行。
listener.beforeJob(execution);
这里就是监听器的前处理器,我们回头找一下会有哪一些监听器会注入。这个监听器主要有俩个方法beforeJob(JobExecution jobExecution)
和afterJob(JobExecution jobExecution);
。
主要是SimpleJobBuilder#build()
->super.enhance(job);
->job.setJobExecutionListeners(listeners.toArray(new JobExecutionListener[0]));
这一段加入的。也就是JobBuilderHelper#properties#jobExecutionListeners
这个属性里面的值。
也就是我们如果有jobBuilders.get("test").listener(new MyJobExecutionListener)
。就会创建一个监听器。
还有一种方法:
JobBuilderHelper#listener(Object listener)
/**
* Registers objects using the annotation based listener configuration.
*
* @param listener the object that has a method configured with listener annotation
* @return this for fluent chaining
*/
public B listener(Object listener) {
Set<Method> jobExecutionListenerMethods = new HashSet<>();
jobExecutionListenerMethods.addAll(ReflectionUtils.findMethod(listener.getClass(), BeforeJob.class));
jobExecutionListenerMethods.addAll(ReflectionUtils.findMethod(listener.getClass(), AfterJob.class));
if (jobExecutionListenerMethods.size() > 0) {
JobListenerFactoryBean factory = new JobListenerFactoryBean();
factory.setDelegate(listener);
properties.addJobExecutionListener((JobExecutionListener) factory.getObject());
}
@SuppressWarnings("unchecked")
B result = (B) this;
return result;
}
参数是一个 object,里面只要有BeforeJob.class
或者AfterJob.class
注解也是JobExecutionListener
监听器。现在看这个名字就能理解为什么是JobExecution
了。
接着回到AbstractJob#execute(JobExecution execution)
doExecute(execution);
@Override
protected void doExecute(JobExecution execution) throws JobInterruptedException, JobRestartException,
StartLimitExceededException {
StepExecution stepExecution = null;
for (Step step : steps) {
stepExecution = handleStep(step, execution);
if (stepExecution.getStatus() != BatchStatus.COMPLETED) {
break;
}
}
//
// Update the job status to be the same as the last step
//
if (stepExecution != null) {
if (logger.isDebugEnabled()) {
logger.debug("Upgrading JobExecution status: " + stepExecution);
}
execution.upgradeStatus(stepExecution.getStatus());
execution.setExitStatus(stepExecution.getExitStatus());
}
}
循环所有的steps
。
stepExecution = handleStep(step, execution);
protected final StepExecution handleStep(Step step, JobExecution execution)
throws JobInterruptedException, JobRestartException,
StartLimitExceededException {
return stepHandler.handleStep(step, execution);
}
这里的stepHandler
对象,是AbstractJob#setJobRepository(JobRepository jobRepository)
的时候创建的。代码如下
public void setJobRepository(JobRepository jobRepository) {
this.jobRepository = jobRepository;
stepHandler = new SimpleStepHandler(jobRepository);
}
SimpleStepHandler#handleStep(Step step, JobExecution execution)
接着到了SimpleStepHandler#handleStep(Step step, JobExecution execution)
@Override
public StepExecution handleStep(Step step, JobExecution execution) throws JobInterruptedException,
JobRestartException, StartLimitExceededException {
if (execution.isStopping()) {
throw new JobInterruptedException("JobExecution interrupted.");
}
JobInstance jobInstance = execution.getJobInstance();
StepExecution lastStepExecution = jobRepository.getLastStepExecution(jobInstance, step.getName());
if (stepExecutionPartOfExistingJobExecution(execution, lastStepExecution)) {
lastStepExecution = null;
}
StepExecution currentStepExecution = lastStepExecution;
if (shouldStart(lastStepExecution, execution, step)) {
currentStepExecution = execution.createStepExecution(step.getName());
boolean isRestart = (lastStepExecution != null && !lastStepExecution.getStatus().equals(
BatchStatus.COMPLETED));
if (isRestart) {
currentStepExecution.setExecutionContext(lastStepExecution.getExecutionContext());
if (lastStepExecution.getExecutionContext().containsKey("batch.executed")) {
currentStepExecution.getExecutionContext().remove("batch.executed");
}
} else {
currentStepExecution.setExecutionContext(new ExecutionContext(executionContext));
}
jobRepository.add(currentStepExecution);
try {
step.execute(currentStepExecution);
currentStepExecution.getExecutionContext().put("batch.executed", true);
} catch (JobInterruptedException e) {
execution.setStatus(BatchStatus.STOPPING);
throw e;
}
jobRepository.updateExecutionContext(execution);
if (currentStepExecution.getStatus() == BatchStatus.STOPPING
|| currentStepExecution.getStatus() == BatchStatus.STOPPED) {
// Ensure that the job gets the message that it is stopping
execution.setStatus(BatchStatus.STOPPING);
throw new JobInterruptedException("Job interrupted by step execution");
}
}
return currentStepExecution;
}
首先获取JobInstance
对象,是MapJobInstanceDao
对象创建的。
StepExecution lastStepExecution = jobRepository.getLastStepExecution(jobInstance, step.getName());
SimpleJobRepository#getLastStepExecution(JobInstance jobInstance, String stepName)
@Override
@Nullable
public StepExecution getLastStepExecution(JobInstance jobInstance, String stepName) {
StepExecution latest = stepExecutionDao.getLastStepExecution(jobInstance, stepName);
if (latest != null) {
ExecutionContext stepExecutionContext = ecDao.getExecutionContext(latest);
latest.setExecutionContext(stepExecutionContext);
ExecutionContext jobExecutionContext = ecDao.getExecutionContext(latest.getJobExecution());
latest.getJobExecution().setExecutionContext(jobExecutionContext);
}
return latest;
}
进入MapStepExecutionDao#getLastStepExecution(JobInstance jobInstance, String stepName)
@Override
public StepExecution getLastStepExecution(JobInstance jobInstance, String stepName) {
StepExecution latest = null;
for (StepExecution stepExecution : executionsByStepExecutionId.values()) {
if (!stepExecution.getStepName().equals(stepName)
|| stepExecution.getJobExecution().getJobInstance().getInstanceId() != jobInstance.getInstanceId()) {
continue;
}
if (latest == null) {
latest = stepExecution;
}
if (latest.getStartTime().getTime() < stepExecution.getStartTime().getTime()) {
latest = stepExecution;
}
if (latest.getStartTime().getTime() == stepExecution.getStartTime().getTime() &&
latest.getId() < stepExecution.getId()) {
latest = stepExecution;
}
}
return latest;
}
第一次进来这里是没有值的,直接返回 null。
回到SimpleStepHandler#handleStep(Step step, JobExecution execution)
stepExecutionPartOfExistingJobExecution(execution, lastStepExecution)
为 false
接着shouldStart(lastStepExecution, execution, step)
SimpleStepHandler#shouldStart(StepExecution lastStepExecution, JobExecution jobExecution, Step step)
// 给定一个步骤和配置,如果该步骤应该开始,则返回 true;如果不应该启动,则返回 false;如果作业应该完成,则抛出异常。
protected boolean shouldStart(StepExecution lastStepExecution, JobExecution jobExecution, Step step)
throws JobRestartException, StartLimitExceededException {
BatchStatus stepStatus;
if (lastStepExecution == null) {
stepStatus = BatchStatus.STARTING;
} else {
stepStatus = lastStepExecution.getStatus();
}
if (stepStatus == BatchStatus.UNKNOWN) {
throw new JobRestartException("Cannot restart step from UNKNOWN status. "
+ "The last execution ended with a failure that could not be rolled back, "
+ "so it may be dangerous to proceed. Manual intervention is probably necessary.");
}
if ((stepStatus == BatchStatus.COMPLETED && !step.isAllowStartIfComplete())
|| stepStatus == BatchStatus.ABANDONED) {
// step is complete, false should be returned, indicating that the
// step should not be started
if (logger.isInfoEnabled()) {
logger.info("Step already complete or not restartable, so no action to execute: " + lastStepExecution);
}
return false;
}
if (jobRepository.getStepExecutionCount(jobExecution.getJobInstance(), step.getName()) < step.getStartLimit()) {
// step start count is less than start max, return true
return true;
} else {
// start max has been exceeded, throw an exception.
throw new StartLimitExceededException("Maximum start limit exceeded for step: " + step.getName()
+ "StartMax: " + step.getStartLimit());
}
}
lastStepExecution
为空,stepStatus
设置为BatchStatus.STARTING
。
step.isAllowStartIfComplete()
是否能够重复执行
@Bean
public Step step() {
TaskletStep taskletStep = stepBuilders.get("test-step")
.<Integer, Integer>chunk(2)
.reader(itemReader())
.writer(itemWriter())
.startLimit(2)
.build();
taskletStep.setAllowStartIfComplete(true);
return taskletStep;
}
改一下,设置成重复执行。taskletStep.setAllowStartIfComplete(true);
,如果不能重复执行打印日志Step already complete or not restartable, so no action to execute
。
jobRepository.getStepExecutionCount(jobExecution.getJobInstance(), step.getName()) < step.getStartLimit()
。
startLimit(2)
。最多允许执行几次,每执行一次,StepExecutionDao#executionsByStepExecutionId
中的数据就会多一条。根据startLimit
判断是否还能重复执行。
继续执行会报错org.springframework.batch.core.StartLimitExceededException: Maximum start limit exceeded for step: test-stepStartMax: 2
currentStepExecution = execution.createStepExecution(step.getName());
public StepExecution createStepExecution(String stepName) {
StepExecution stepExecution = new StepExecution(stepName, this);
this.stepExecutions.add(stepExecution);
return stepExecution;
}
创建一个StepExecution
并返回。stepExecutions.add
boolean isRestart = (lastStepExecution != null && !lastStepExecution.getStatus().equals(BatchStatus.COMPLETED));
lastStepExecution
为空。进入 else。
currentStepExecution.setExecutionContext(new ExecutionContext(executionContext));
普通 set,executionContext
也是一个普通的new ExecutionContext()
。构造器方法中 new 的。
jobRepository.add(currentStepExecution);
SimpleJobRepository#add(StepExecution stepExecution)
@Override
public void add(StepExecution stepExecution) {
validateStepExecution(stepExecution);
stepExecution.setLastUpdated(new Date(System.currentTimeMillis()));
stepExecutionDao.saveStepExecution(stepExecution);
ecDao.saveExecutionContext(stepExecution);
}
validateStepExecution(stepExecution)
:校验
stepExecution.setLastUpdated(new Date(System.currentTimeMillis()));
:设置最后修改时间
stepExecutionDao.saveStepExecution(stepExecution);
@Override
public void saveStepExecution(StepExecution stepExecution) {
Assert.isTrue(stepExecution.getId() == null, "stepExecution id was not null");
Assert.isTrue(stepExecution.getVersion() == null, "stepExecution version was not null");
Assert.notNull(stepExecution.getJobExecutionId(), "JobExecution must be saved already.");
Map<Long, StepExecution> executions = executionsByJobExecutionId.get(stepExecution.getJobExecutionId());
if (executions == null) {
executions = new ConcurrentHashMap<>();
executionsByJobExecutionId.put(stepExecution.getJobExecutionId(), executions);
}
stepExecution.setId(currentId.incrementAndGet());
stepExecution.incrementVersion();
StepExecution copy = copy(stepExecution);
executions.put(stepExecution.getId(), copy);
executionsByStepExecutionId.put(stepExecution.getId(), copy);
}
先校验,再新增一条数据。 StepExecution
里面包含了JobExecution
。 JobExecutionDao
维护jobExecution
的 Id。StepExecution
维护StepExecution
的 Id。
ecDao.saveExecutionContext(stepExecution);
@Override
public void saveExecutionContext(StepExecution stepExecution) {
updateExecutionContext(stepExecution);
}
@Override
public void updateExecutionContext(StepExecution stepExecution) {
ExecutionContext executionContext = stepExecution.getExecutionContext();
if (executionContext != null) {
contexts.put(ContextKey.step(stepExecution.getId()), copy(executionContext));
}
}
新增一条数据。前面已经新增了一条job
。现在新增的是step
。
step.execute(currentStepExecution);
继续回到SimpleStepHandler#handleStep(Step step, JobExecution execution)
。
AbstractStep#execute(StepExecution stepExecution)
/**
* Template method for step execution logic - calls abstract methods for resource initialization (
* {@link #open(ExecutionContext)}), execution logic ({@link #doExecute(StepExecution)}) and resource closing (
* {@link #close(ExecutionContext)}).
*/
@Override
public final void execute(StepExecution stepExecution) throws JobInterruptedException,
UnexpectedJobExecutionException {
Assert.notNull(stepExecution, "stepExecution must not be null");
if (logger.isDebugEnabled()) {
logger.debug("Executing: id=" + stepExecution.getId());
}
stepExecution.setStartTime(new Date());
stepExecution.setStatus(BatchStatus.STARTED);
Timer.Sample sample = BatchMetrics.createTimerSample();
getJobRepository().update(stepExecution);
// Start with a default value that will be trumped by anything
ExitStatus exitStatus = ExitStatus.EXECUTING;
doExecutionRegistration(stepExecution);
try {
getCompositeListener().beforeStep(stepExecution);
open(stepExecution.getExecutionContext());
try {
doExecute(stepExecution);
} catch (RepeatException e) {
throw e.getCause();
}
exitStatus = ExitStatus.COMPLETED.and(stepExecution.getExitStatus());
// Check if someone is trying to stop us
if (stepExecution.isTerminateOnly()) {
throw new JobInterruptedException("JobExecution interrupted.");
}
// Need to upgrade here not set, in case the execution was stopped
stepExecution.upgradeStatus(BatchStatus.COMPLETED);
if (logger.isDebugEnabled()) {
logger.debug("Step execution success: id=" + stepExecution.getId());
}
} catch (Throwable e) {
stepExecution.upgradeStatus(determineBatchStatus(e));
exitStatus = exitStatus.and(getDefaultExitStatusForFailure(e));
stepExecution.addFailureException(e);
if (stepExecution.getStatus() == BatchStatus.STOPPED) {
logger.info(String.format("Encountered interruption executing step %s in job %s : %s", name, stepExecution.getJobExecution().getJobInstance().getJobName(), e.getMessage()));
if (logger.isDebugEnabled()) {
logger.debug("Full exception", e);
}
} else {
logger.error(String.format("Encountered an error executing step %s in job %s", name, stepExecution.getJobExecution().getJobInstance().getJobName()), e);
}
} finally {
try {
// Update the step execution to the latest known value so the
// listeners can act on it
exitStatus = exitStatus.and(stepExecution.getExitStatus());
stepExecution.setExitStatus(exitStatus);
exitStatus = exitStatus.and(getCompositeListener().afterStep(stepExecution));
} catch (Exception e) {
logger.error(String.format("Exception in afterStep callback in step %s in job %s", name, stepExecution.getJobExecution().getJobInstance().getJobName()), e);
}
try {
getJobRepository().updateExecutionContext(stepExecution);
} catch (Exception e) {
stepExecution.setStatus(BatchStatus.UNKNOWN);
exitStatus = exitStatus.and(ExitStatus.UNKNOWN);
stepExecution.addFailureException(e);
logger.error(String.format("Encountered an error saving batch meta data for step %s in job %s. "
+ "This job is now in an unknown state and should not be restarted.", name, stepExecution.getJobExecution().getJobInstance().getJobName()), e);
}
sample.stop(BatchMetrics.createTimer("step", "Step duration",
Tag.of("job.name", stepExecution.getJobExecution().getJobInstance().getJobName()),
Tag.of("name", stepExecution.getStepName()),
Tag.of("status", stepExecution.getExitStatus().getExitCode())
));
stepExecution.setEndTime(new Date());
stepExecution.setExitStatus(exitStatus);
Duration stepExecutionDuration = BatchMetrics.calculateDuration(stepExecution.getStartTime(), stepExecution.getEndTime());
if (logger.isInfoEnabled()) {
logger.info("Step: [" + stepExecution.getStepName() + "] executed in " + BatchMetrics.formatDuration(stepExecutionDuration));
}
try {
getJobRepository().update(stepExecution);
} catch (Exception e) {
stepExecution.setStatus(BatchStatus.UNKNOWN);
stepExecution.setExitStatus(exitStatus.and(ExitStatus.UNKNOWN));
stepExecution.addFailureException(e);
logger.error(String.format("Encountered an error saving batch meta data for step %s in job %s. "
+ "This job is now in an unknown state and should not be restarted.", name, stepExecution.getJobExecution().getJobInstance().getJobName()), e);
}
try {
close(stepExecution.getExecutionContext());
} catch (Exception e) {
logger.error(String.format("Exception while closing step execution resources in step %s in job %s", name, stepExecution.getJobExecution().getJobInstance().getJobName()), e);
stepExecution.addFailureException(e);
}
doExecutionRelease();
if (logger.isDebugEnabled()) {
logger.debug("Step execution complete: " + stepExecution.getSummary());
}
}
}
前面几行校验,打印日志,计时器。
getJobRepository().update(stepExecution);
@Override
public void update(StepExecution stepExecution) {
validateStepExecution(stepExecution);
Assert.notNull(stepExecution.getId(), "StepExecution must already be saved (have an id assigned)");
stepExecution.setLastUpdated(new Date(System.currentTimeMillis()));
stepExecutionDao.updateStepExecution(stepExecution);
checkForInterruption(stepExecution);
}
校验参数,设置最后修改时间。
stepExecutionDao.updateStepExecution(stepExecution);
跟前面一样,版本号升级,新增一条数据。这里有俩个 Map,executionsByStepExecutionId
和executionsByJobExecutionId
executionsByJobExecutionId
:
private Map<Long, Map<Long, StepExecution>> executionsByJobExecutionId = new ConcurrentHashMap<>();
key 为Job
的 id。然后job
有多个Step
。泛型里面的 map 就是Step
的 id 和对象。
executionsByStepExecutionId
:
private Map<Long, StepExecution> executionsByStepExecutionId = new ConcurrentHashMap<>();
key 为 Step 的 id
checkForInterruption(stepExecution);
回到SimpleJobRepository#update(StepExecution stepExecution)
。
/**
* Check to determine whether or not the JobExecution that is the parent of
* the provided StepExecution has been interrupted. If, after synchronizing
* the status with the database, the status has been updated to STOPPING,
* then the job has been interrupted.
*
* @param stepExecution
*/
private void checkForInterruption(StepExecution stepExecution) {
JobExecution jobExecution = stepExecution.getJobExecution();
jobExecutionDao.synchronizeStatus(jobExecution);
if (jobExecution.isStopping()) {
logger.info("Parent JobExecution is stopped, so passing message on to StepExecution");
stepExecution.setTerminateOnly();
}
}
检查以确定作为所提供的 StepExecution 的父级 JobExecution 是否已被中断。如果与数据库同步状态后,状态已更新为
STOPPING,则作业已中断。
jobExecutionDao.synchronizeStatus(jobExecution);
:版本号是否改变了。
ExitStatus exitStatus = ExitStatus.EXECUTING;
回到AbstractStep#execute(StepExecution stepExecution)
方法,创建exitStatus
doExecutionRegistration(stepExecution);
前面是JobSynchronizationManager.register(execution);
。注入的是 job,现在这个是 step。逻辑一样的。
getCompositeListener().beforeStep(stepExecution);
跟 job 也基本一致,在 taskletStep = stepBuilders.get("test-step").listener(new StepExecutionListener() {
中或者listener(Object listener)
里面存在BeforeStep.class
和AfterStep.class
注解的方法。 可以得出xxExecutionListener
是比较前面执行的。
open(stepExecution.getExecutionContext());
@Override
protected void open(ExecutionContext ctx) throws Exception {
stream.open(ctx);
}
里面的stream
存在一个private final List<ItemStream> streams = new ArrayList<>();
属性,也就是ItemStream
对象。里面的
open 方法会在这里执行。
doExecute(stepExecution);
@Override
protected void doExecute(StepExecution stepExecution) throws Exception {
stepExecution.getExecutionContext().put(TASKLET_TYPE_KEY, tasklet.getClass().getName());
stepExecution.getExecutionContext().put(STEP_TYPE_KEY, this.getClass().getName());
stream.update(stepExecution.getExecutionContext());
getJobRepository().updateExecutionContext(stepExecution);
final Semaphore semaphore = createSemaphore();
stepOperations.iterate(new StepContextRepeatCallback(stepExecution) {
@Override
public RepeatStatus doInChunkContext(RepeatContext repeatContext, ChunkContext chunkContext)
throws Exception {
StepExecution stepExecution = chunkContext.getStepContext().getStepExecution();
interruptionPolicy.checkInterrupted(stepExecution);
RepeatStatus result;
try {
result = new TransactionTemplate(transactionManager, transactionAttribute)
.execute(new ChunkTransactionCallback(chunkContext, semaphore));
} catch (UncheckedTransactionException e) {
throw (Exception) e.getCause();
}
chunkListener.afterChunk(chunkContext);
interruptionPolicy.checkInterrupted(stepExecution);
return result == null ? RepeatStatus.FINISHED : result;
}
});
}
首先在stepExecution#ExecutionContext
set 了batch.taskletType
和batch.stepType
。接着执行了stream.update(stepExecution.getExecutionContext());
。与上面一样,ItemStream#update
方法都在这里执行的。
又是更新getJobRepository().updateExecutionContext(stepExecution);
。与上面一样。里面主要就是环境更新ecDao.updateExecutionContext(stepExecution);
。
final Semaphore semaphore = createSemaphore();
每个步骤执行共享信号量,因此其他步骤执行可以并行运行而不需要锁,这里只是 new 一个对象,不是使用。
stepOperations.iterate(new StepContextRepeatCallback(stepExecution)
stepOperations
是一个RepeatTemplate
对象,实现了RepeatOperations
。里面只有一个方法。RepeatStatus#iterate(RepeatCallback callback)
。
RepeatTemplate#iterate(RepeatCallback callback)
/**
* Execute the batch callback until the completion policy decides that we
* are finished. Wait for the whole batch to finish before returning even if
* the task executor is asynchronous.
*
* @see org.springframework.batch.repeat.RepeatOperations#iterate(org.springframework.batch.repeat.RepeatCallback)
*/
@Override
public RepeatStatus iterate(RepeatCallback callback) {
RepeatContext outer = RepeatSynchronizationManager.getContext();
RepeatStatus result = RepeatStatus.CONTINUABLE;
try {
// This works with an asynchronous TaskExecutor: the
// interceptors have to wait for the child processes.
result = executeInternal(callback);
} finally {
RepeatSynchronizationManager.clear();
if (outer != null) {
RepeatSynchronizationManager.register(outer);
}
}
return result;
}
第一次进来outer
为空,进入result = executeInternal(callback);
方法。
private RepeatStatus executeInternal(final RepeatCallback callback) {
// Reset the termination policy if there is one...
RepeatContext context = start();
// Make sure if we are already marked complete before we start then no
// processing takes place.
boolean running = !isMarkedComplete(context);
for (RepeatListener interceptor : listeners) {
interceptor.open(context);
running = running && !isMarkedComplete(context);
if (!running)
break;
}
// Return value, default is to allow continued processing.
RepeatStatus result = RepeatStatus.CONTINUABLE;
RepeatInternalState state = createInternalState(context);
// This is the list of exceptions thrown by all active callbacks
Collection<Throwable> throwables = state.getThrowables();
// Keep a separate list of exceptions we handled that need to be
// rethrown
Collection<Throwable> deferred = new ArrayList<>();
try {
while (running) {
/*
* Run the before interceptors here, not in the task executor so
* that they all happen in the same thread - it's easier for
* tracking batch status, amongst other things.
*/
for (int i = 0; i < listeners.length; i++) {
RepeatListener interceptor = listeners[i];
interceptor.before(context);
// Allow before interceptors to veto the batch by setting
// flag.
running = running && !isMarkedComplete(context);
}
// Check that we are still running (should always be true) ...
if (running) {
try {
result = getNextResult(context, callback, state);
executeAfterInterceptors(context, result);
} catch (Throwable throwable) {
doHandle(throwable, context, deferred);
}
// N.B. the order may be important here:
if (isComplete(context, result) || isMarkedComplete(context) || !deferred.isEmpty()) {
running = false;
}
}
}
result = result.and(waitForResults(state));
for (Throwable throwable : throwables) {
doHandle(throwable, context, deferred);
}
// Explicitly drop any references to internal state...
state = null;
}
/*
* No need for explicit catch here - if the business processing threw an
* exception it was already handled by the helper methods. An exception
* here is necessarily fatal.
*/ finally {
try {
if (!deferred.isEmpty()) {
Throwable throwable = deferred.iterator().next();
if (logger.isDebugEnabled()) {
logger.debug("Handling fatal exception explicitly (rethrowing first of " + deferred.size() + "): "
+ throwable.getClass().getName() + ": " + throwable.getMessage());
}
rethrow(throwable);
}
} finally {
try {
for (int i = listeners.length; i-- > 0; ) {
RepeatListener interceptor = listeners[i];
interceptor.close(context);
}
} finally {
context.close();
}
}
}
return result;
}
RepeatContext context = start();
/**
* Delegate to the {@link CompletionPolicy}.
*
* @return a {@link RepeatContext} object that can be used by the implementation to store
* internal state for a batch step.
* @see org.springframework.batch.repeat.CompletionPolicy#start(RepeatContext)
*/
protected RepeatContext start() {
RepeatContext parent = RepeatSynchronizationManager.getContext();
RepeatContext context = completionPolicy.start(parent);
RepeatSynchronizationManager.register(context);
logger.debug("Starting repeat context.");
return context;
}
往RepeatSynchronizationManager#ThreadLocal<RepeatContext> contextHolder = new ThreadLocal<>()
中新增一个RepeatContext
。这里新增的是一个RepeatContextSupport
对象。
boolean running = !isMarkedComplete(context);
确保我们在开始之前是否已标记为完成,则不会进行任何处理。
private boolean isMarkedComplete(RepeatContext context) {
boolean complete = context.isCompleteOnly();
if (context.getParent() != null) {
complete = complete || isMarkedComplete(context.getParent());
}
if (complete) {
logger.debug("Repeat is complete according to context alone.");
}
return complete;
}
complete
默认为 false,取反running
为 true。
listeners
这个对象是在 builder 中赋值的,我们没有进行别的处理。
RepeatStatus result = RepeatStatus.CONTINUABLE;
返回值,默认是允许继续处理。
RepeatInternalState state = createInternalState(context);
protected RepeatInternalState createInternalState(RepeatContext context) {
return new RepeatInternalStateSupport();
}
Collection throwables = state.getThrowables();
这是所有活动回调抛出的异常列表
Collection deferred = new ArrayList<>();
保留我们处理过的需要重新抛出的异常的单独列表
接下来进入 while 循环,running
为 true,第一个 for 循环listeners
值为空,跳过。
判断 if 进入。
result = getNextResult(context, callback, state);
protected RepeatStatus getNextResult(RepeatContext context, RepeatCallback callback, RepeatInternalState state)
throws Throwable {
update(context);
if (logger.isDebugEnabled()) {
logger.debug("Repeat operation about to start at count=" + context.getStartedCount());
}
return callback.doInIteration(context);
}
解释一下三个参数,context
,前面方法start()
中 new 出来的RepeatContextSupport
,当前里面什么都没有。
callback
是 new 的一个new StepContextRepeatCallback(stepExecution)
对象。
state
是 new RepeatInternalStateSupport()
对象。在前面调用createInternalState(context)
创建出来的。
update(context);
更新context
,这里更新就是把RepeatContextSupport#count
++,次数加一。
return callback.doInIteration(context);
StepContextRepeatCallback#doInIteration(RepeatContext context)
@Override
public RepeatStatus doInIteration(RepeatContext context) throws Exception {
// The StepContext has to be the same for all chunks,
// otherwise step-scoped beans will be re-initialised for each chunk.
StepContext stepContext = StepSynchronizationManager.register(stepExecution);
if (logger.isDebugEnabled()) {
logger.debug("Preparing chunk execution for StepContext: " + ObjectUtils.identityToString(stepContext));
}
ChunkContext chunkContext = attributeQueue.poll();
if (chunkContext == null) {
chunkContext = new ChunkContext(stepContext);
}
try {
if (logger.isDebugEnabled()) {
logger.debug("Chunk execution starting: queue size=" + attributeQueue.size());
}
return doInChunkContext(context, chunkContext);
} finally {
// Still some stuff to do with the data in this chunk,
// pass it back.
if (!chunkContext.isComplete()) {
attributeQueue.add(chunkContext);
}
StepSynchronizationManager.close();
}
}
StepContext stepContext = StepSynchronizationManager.register(stepExecution);
简单来说就是SynchronizationManagerSupport#counts
++。
接着 chunkContext = new ChunkContext(stepContext);
doInChunkContext(context, chunkContext);
@Override
public RepeatStatus doInChunkContext(RepeatContext repeatContext, ChunkContext chunkContext)
throws Exception {
StepExecution stepExecution = chunkContext.getStepContext().getStepExecution();
// Before starting a new transaction, check for
// interruption.
interruptionPolicy.checkInterrupted(stepExecution);
RepeatStatus result;
try {
result = new TransactionTemplate(transactionManager, transactionAttribute)
.execute(new ChunkTransactionCallback(chunkContext, semaphore));
} catch (UncheckedTransactionException e) {
// Allow checked exceptions to be thrown inside callback
throw (Exception) e.getCause();
}
chunkListener.afterChunk(chunkContext);
// Check for interruption after transaction as well, so that
// the interrupted exception is correctly propagated up to
// caller
interruptionPolicy.checkInterrupted(stepExecution);
return result == null ? RepeatStatus.FINISHED : result;
}
执行 doInChunkContext(RepeatContext repeatContext, ChunkContext chunkContext)
。这个是在TaskletStep#doExecute(StepExecution stepExecution)
中,stepOperations.iterate(new StepContextRepeatCallback(stepExecution) {
里面的doInChunkContext(RepeatContext repeatContext, ChunkContext chunkContext)
方法。
StepExecution stepExecution = chunkContext.getStepContext().getStepExecution();
获取stepExecution
,StepContext
里面存有StepExecution
。
interruptionPolicy.checkInterrupted(stepExecution);
在开始新事务之前,检查是否有中断。
result = new TransactionTemplate(transactionManager, transactionAttribute)
创建一个TransactionTemplate
。transactionManager
是默认的ResourcelessTransactionManager
。
transactionAttribute
是new DefaultTransactionAttribute(
。
.execute(new ChunkTransactionCallback(chunkContext, semaphore));
执行 execute 方法。使用ChunkTransactionCallback
包装chunkContext, semaphore
俩个属性。
@Override
@Nullable
public <T> T execute(TransactionCallback<T> action) throws TransactionException {
Assert.state(this.transactionManager != null, "No PlatformTransactionManager set");
if (this.transactionManager instanceof CallbackPreferringPlatformTransactionManager) {
return ((CallbackPreferringPlatformTransactionManager) this.transactionManager).execute(this, action);
} else {
TransactionStatus status = this.transactionManager.getTransaction(this);
T result;
try {
result = action.doInTransaction(status);
} catch (RuntimeException | Error ex) {
// Transactional code threw application exception -> rollback
rollbackOnException(status, ex);
throw ex;
} catch (Throwable ex) {
// Transactional code threw unexpected exception -> rollback
rollbackOnException(status, ex);
throw new UndeclaredThrowableException(ex, "TransactionCallback threw undeclared checked exception");
}
this.transactionManager.commit(status);
return result;
}
}
进入 else,里面大概逻辑就是成功就提交,失败就回滚。主要方法在result = action.doInTransaction(status);
中。
result = action.doInTransaction(status);
TransactionCallback<T> action
对象为ChunkTransactionCallback
@Override
public RepeatStatus doInTransaction(TransactionStatus status) {
TransactionSynchronizationManager.registerSynchronization(this);
RepeatStatus result = RepeatStatus.CONTINUABLE;
StepContribution contribution = stepExecution.createStepContribution();
chunkListener.beforeChunk(chunkContext);
// In case we need to push it back to its old value
// after a commit fails...
oldVersion = new StepExecution(stepExecution.getStepName(), stepExecution.getJobExecution());
copy(stepExecution, oldVersion);
try {
try {
try {
result = tasklet.execute(contribution, chunkContext);
if (result == null) {
result = RepeatStatus.FINISHED;
}
} catch (Exception e) {
if (transactionAttribute.rollbackOn(e)) {
chunkContext.setAttribute(ChunkListener.ROLLBACK_EXCEPTION_KEY, e);
throw e;
}
}
} finally {
// If the step operations are asynchronous then we need
// to synchronize changes to the step execution (at a
// minimum). Take the lock *before* changing the step
// execution.
try {
semaphore.acquire();
locked = true;
} catch (InterruptedException e) {
logger.error("Thread interrupted while locking for repository update");
stepExecution.setStatus(BatchStatus.STOPPED);
stepExecution.setTerminateOnly();
Thread.currentThread().interrupt();
}
// Apply the contribution to the step
// even if unsuccessful
if (logger.isDebugEnabled()) {
logger.debug("Applying contribution: " + contribution);
}
stepExecution.apply(contribution);
}
stepExecutionUpdated = true;
stream.update(stepExecution.getExecutionContext());
try {
// Going to attempt a commit. If it fails this flag will
// stay false and we can use that later.
getJobRepository().updateExecutionContext(stepExecution);
stepExecution.incrementCommitCount();
if (logger.isDebugEnabled()) {
logger.debug("Saving step execution before commit: " + stepExecution);
}
getJobRepository().update(stepExecution);
} catch (Exception e) {
// If we get to here there was a problem saving the step
// execution and we have to fail.
String msg = "JobRepository failure forcing rollback";
logger.error(msg, e);
throw new FatalStepExecutionException(msg, e);
}
} catch (Error e) {
if (logger.isDebugEnabled()) {
logger.debug("Rollback for Error: " + e.getClass().getName() + ": " + e.getMessage());
}
rollback(stepExecution);
throw e;
} catch (RuntimeException e) {
if (logger.isDebugEnabled()) {
logger.debug("Rollback for RuntimeException: " + e.getClass().getName() + ": " + e.getMessage());
}
rollback(stepExecution);
throw e;
} catch (Exception e) {
if (logger.isDebugEnabled()) {
logger.debug("Rollback for Exception: " + e.getClass().getName() + ": " + e.getMessage());
}
rollback(stepExecution);
// Allow checked exceptions
throw new UncheckedTransactionException(e);
}
return result;
}
TransactionSynchronizationManager.registerSynchronization(this);
为当前线程注册一个新的事务同步。通常由资源管理代码调用。
RepeatStatus result = RepeatStatus.CONTINUABLE;
默认返回。
StepContribution contribution = stepExecution.createStepContribution();
创建一个StepContribution
。return new StepContribution(this);
chunkListener.beforeChunk(chunkContext);
里面的chunkListener
前处理器。 也是在 builder 中 set 的。与监听器逻辑一样。打开ChunkListener
类,把注释翻译一下,对照这里就知道了。
oldVersion = new StepExecution(stepExecution.getStepName(), stepExecution.getJobExecution()); copy(stepExecution, oldVersion);
如果提交失败后我们需要将其推回到旧值......,赋值一份用来进行回滚。
result = tasklet.execute(contribution, chunkContext);
tasklet
是SimpleStepBuilder#createTasklet()
里面创建的。也就是ChunkOrientedTasklet
。
public RepeatStatus execute(StepContribution contribution, ChunkContext chunkContext) throws Exception {
@SuppressWarnings("unchecked")
Chunk<I> inputs = (Chunk<I>) chunkContext.getAttribute(INPUTS_KEY);
if (inputs == null) {
inputs = chunkProvider.provide(contribution);
if (buffering) {
chunkContext.setAttribute(INPUTS_KEY, inputs);
}
}
chunkProcessor.process(contribution, inputs);
chunkProvider.postProcess(contribution, inputs);
// Allow a message coming back from the processor to say that we
// are not done yet
if (inputs.isBusy()) {
logger.debug("Inputs still busy");
return RepeatStatus.CONTINUABLE;
}
chunkContext.removeAttribute(INPUTS_KEY);
chunkContext.setComplete();
if (logger.isDebugEnabled()) {
logger.debug("Inputs not busy, ended: " + inputs.isEnd());
}
return RepeatStatus.continueIf(!inputs.isEnd());
}
这里又是使用repeatOperations#.iterate(new RepeatCallback()
。这里有个点需要注意一下,在 builder
中。存在俩个RepeatOperations
对象。一个是SimpleStepBuilder#RepeatOperations chunkOperations
。名字中有个chunk
的。还有一个在父类AbstractTaskletStepBuilder#RepeatOperations stepOperations
中。
chunkOperations
在这里使用的,父类的在TaskletStep#doExecute(StepExecution stepExecution)
中使用的。
@Override
public Chunk<I> provide(final StepContribution contribution) throws Exception {
final Chunk<I> inputs = new Chunk<>();
repeatOperations.iterate(new RepeatCallback() {
@Override
public RepeatStatus doInIteration(final RepeatContext context) throws Exception {
I item = null;
Timer.Sample sample = Timer.start(Metrics.globalRegistry);
String status = BatchMetrics.STATUS_SUCCESS;
try {
item = read(contribution, inputs);
} catch (SkipOverflowException e) {
// read() tells us about an excess of skips by throwing an
// exception
status = BatchMetrics.STATUS_FAILURE;
return RepeatStatus.FINISHED;
} finally {
stopTimer(sample, contribution.getStepExecution(), status);
}
if (item == null) {
inputs.setEnd();
return RepeatStatus.FINISHED;
}
inputs.add(item);
contribution.incrementReadCount();
return RepeatStatus.CONTINUABLE;
}
});
return inputs;
}
这里的repeatOperations
也是RepeatTemplate
。逻辑与前面的都一致,只不过RepeatCallback
不同,直接进入RepeatCallback#doInIteration(final RepeatContext context)
方法。
chunkProvider.provide(contribution);
这个就是读取数据接口了。先看看参数。StepContribution
里面只有一个stepExecution
。chunkProvider
是SimpleChunkProvider<I> chunkProvider = new SimpleChunkProvider<>(getReader(), repeatOperations);
这样的,repeatOperations
是一个RepeatTemplate
。
进入方法。
@Override
public Chunk<I> provide(final StepContribution contribution) throws Exception {
final Chunk<I> inputs = new Chunk<>();
repeatOperations.iterate(new RepeatCallback() {
@Override
public RepeatStatus doInIteration(final RepeatContext context) throws Exception {
I item = null;
Timer.Sample sample = Timer.start(Metrics.globalRegistry);
String status = BatchMetrics.STATUS_SUCCESS;
try {
item = read(contribution, inputs);
} catch (SkipOverflowException e) {
// read() tells us about an excess of skips by throwing an
// exception
status = BatchMetrics.STATUS_FAILURE;
return RepeatStatus.FINISHED;
} finally {
stopTimer(sample, contribution.getStepExecution(), status);
}
if (item == null) {
inputs.setEnd();
return RepeatStatus.FINISHED;
}
inputs.add(item);
contribution.incrementReadCount();
return RepeatStatus.CONTINUABLE;
}
});
return inputs;
}
只关注一些关键代码,整体逻辑就是执行item = read(contribution, inputs);
,出错返回 RepeatStatus.FINISHED
,正常就把值放入final Chunk<I> inputs = new Chunk<>();
中,并且执行contribution.incrementReadCount();
->
把里面的readCount++
。
关键点:item == null
,就表示后面没有数据了,设置 end 为结束。这就是为什么我们返回 null,就结束,如果不返回 null,就会一直执行下去。
item = read(contribution, inputs);
进入 doRead()
/**
* Surrounds the read call with listener callbacks.
*
* @return the item or {@code null} if the data source is exhausted
* @throws Exception is thrown if error occurs during read.
*/
@Nullable
protected final I doRead() throws Exception {
try {
listener.beforeRead();
I item = itemReader.read();
if (item != null) {
listener.afterRead(item);
}
return item;
} catch (Exception e) {
if (logger.isDebugEnabled()) {
logger.debug(e.getMessage() + " : " + e.getClass().getName());
}
listener.onReadError(e);
throw e;
}
}
listener.beforeRead(); ItemReadListener#beforeRead()
方法。
I item = itemReader.read();
这个就是我们自己写的类了。执行里面的方法,返回一个值。
listener.afterRead(item);
返回值不为空,执行ItemReadListener#afterRead
方法。
接着ChunkOrientedTasklet#execute(StepContribution contribution, ChunkContext chunkContext)
。
chunkProcessor.process(contribution, inputs);
inputs
就是我们自己写的返回的值。contribution
中的 readCount++
。
chunkProcessor
是SimpleChunkProcessor<I, O> chunkProcessor = new SimpleChunkProcessor<>(getProcessor(), getWriter());
@Override
public final void process(StepContribution contribution, Chunk<I> inputs) throws Exception {
// Allow temporary state to be stored in the user data field
initializeUserData(inputs);
// If there is no input we don't have to do anything more
if (isComplete(inputs)) {
return;
}
// Make the transformation, calling remove() on the inputs iterator if
// any items are filtered. Might throw exception and cause rollback.
Chunk<O> outputs = transform(contribution, inputs);
// Adjust the filter count based on available data
contribution.incrementFilterCount(getFilterCount(inputs, outputs));
// Adjust the outputs if necessary for housekeeping purposes, and then
// write them out...
write(contribution, inputs, getAdjustedOutputs(inputs, outputs));
}
initializeUserData(inputs);
允许临时状态存储在用户数据字段中,
protected void initializeUserData(Chunk<I> inputs){
inputs.setUserData(inputs.size());
}
isComplete(inputs)
如果没有输入,我们不需要做任何事情
Chunk outputs = transform(contribution, inputs);
进行转换,如果过滤了任何项目,则在输入迭代器上调用 remove()。可能会抛出异常并导致回滚。
protected Chunk<O> transform(StepContribution contribution, Chunk<I> inputs) throws Exception {
Chunk<O> outputs = new Chunk<>();
for (Chunk<I>.ChunkIterator iterator = inputs.iterator(); iterator.hasNext(); ) {
final I item = iterator.next();
O output;
Timer.Sample sample = BatchMetrics.createTimerSample();
String status = BatchMetrics.STATUS_SUCCESS;
try {
output = doProcess(item);
} catch (Exception e) {
/*
* For a simple chunk processor (no fault tolerance) we are done
* here, so prevent any more processing of these inputs.
*/
inputs.clear();
status = BatchMetrics.STATUS_FAILURE;
throw e;
} finally {
stopTimer(sample, contribution.getStepExecution(), "item.process", status, "Item processing");
}
if (output != null) {
outputs.add(output);
} else {
iterator.remove();
}
}
return outputs;
}
只看一个方法,output = doProcess(item);
doProcess(item)
protected final O doProcess(I item) throws Exception {
if (itemProcessor == null) {
@SuppressWarnings("unchecked")
O result = (O) item;
return result;
}
try {
listener.beforeProcess(item);
O result = itemProcessor.process(item);
listener.afterProcess(item, result);
return result;
} catch (Exception e) {
listener.onProcessError(item, e);
throw e;
}
}
itemProcessor
为空直接返回,如果存在。
先执行listener.beforeProcess(item);
(ItemProcessListener#beforeProcess
)。
然后存在ItemProcessor
类,就执行。这个里面是一个数据转换,我读取数据,不能够直接写入就需要在这里进行一次转换,再进行写入。
最后再执行一次ItemProcessListener#afterProcess
。并返回结果。 如果没有ItemProcessor
,里面的ItemProcessListener
就不会执行。这里需要注意。
接着回到上面。
contribution.incrementFilterCount(getFilterCount(inputs, outputs));
维护里面的filterCount
。看看有没有过滤的数据。
write(contribution, inputs, getAdjustedOutputs(inputs, outputs));
如果需要的话,调整输出以进行内务处理,然后将它们写出来......
/**
* Simple implementation delegates to the {@link #doWrite(List)} method and
* increments the write count in the contribution. Subclasses can handle
* more complicated scenarios, e.g.with fault tolerance. If output items are
* skipped they should be removed from the inputs as well.
*
* @param contribution the current step contribution
* @param inputs the inputs that gave rise to the outputs
* @param outputs the outputs to write
* @throws Exception if there is a problem
*/
protected void write(StepContribution contribution, Chunk<I> inputs, Chunk<O> outputs) throws Exception {
Timer.Sample sample = BatchMetrics.createTimerSample();
String status = BatchMetrics.STATUS_SUCCESS;
try {
doWrite(outputs.getItems());
} catch (Exception e) {
/*
* For a simple chunk processor (no fault tolerance) we are done
* here, so prevent any more processing of these inputs.
*/
inputs.clear();
status = BatchMetrics.STATUS_FAILURE;
throw e;
} finally {
stopTimer(sample, contribution.getStepExecution(), "chunk.write", status, "Chunk writing");
}
contribution.incrementWriteCount(outputs.size());
}
关键俩句,doWrite(outputs.getItems());
和 contribution.incrementWriteCount(outputs.size());
。后面就是添加 WriteCount
。
doWrite(outputs.getItems())
protected final void doWrite(List<O> items) throws Exception {
if (itemWriter == null) {
return;
}
try {
listener.beforeWrite(items);
writeItems(items);
doAfterWrite(items);
} catch (Exception e) {
doOnWriteError(e, items);
throw e;
}
}
listener.beforeWrite(items)
:首先ItemWriteListener#beforeWrite
方法。
writeItems(items);
:自己写的itemWriter
。执行里面的方法。
doAfterWrite(items)
:执行ItemWriteListener#afterWrite
方法。
这里我们要注意,这只是一次,也就是我们写的 stepBuilders.get("test-step").<Integer, Integer>chunk(2)
,读取俩次值放入 Chunk
中,再RepeatTemplate#iterate(RepeatCallback callback)
调用 RepeatTemplate#executeInternal(final RepeatCallback callback)
里面。有个 while
循环,出去条件就是isComplete(context, result) || isMarkedComplete(context) || !deferred.isEmpty()
。这一段我们方法放在外面的RepeatTemplate
中进行讲解。
chunkProvider.postProcess(contribution, inputs);
当前实现类没有处理。
inputs.isBusy()
值为 false。
下面就是删除当前INPUTS_KEY
的值,设置Complete
为 true。
RepeatStatus.continueIf(!inputs.isEnd());
这是很重要的一点,是否已经结束。上面讲到过。是否已经结束了。
又回到TaskletStep#doInTransaction(TransactionStatus status)
,刚刚所有的方法都在tasklet.execute(contribution, chunkContext)
中。
然后接着执行 finally 中的代码。 semaphore.acquire(); 信号量,锁上。 locked = true; 锁=true。
如果步骤操作是异步的,那么我们需要将更改同步到步骤执行(至少)。在更改步骤执行之前锁定。
继续执行stepExecution.apply(contribution);
public synchronized void apply(StepContribution contribution) {
readSkipCount += contribution.getReadSkipCount();
writeSkipCount += contribution.getWriteSkipCount();
processSkipCount += contribution.getProcessSkipCount();
filterCount += contribution.getFilterCount();
readCount += contribution.getReadCount();
writeCount += contribution.getWriteCount();
exitStatus = exitStatus.and(contribution.getExitStatus());
}
相加。
stepExecutionUpdated = true;
等到使用再讲解。
stream.update(stepExecution.getExecutionContext());
现在到了ItemStream#update(ExecutionContext executionContext)
了。
getJobRepository().updateExecutionContext(stepExecution);
MapExecutionContextDao
更新。
getJobRepository().update(stepExecution);
后面详细讲解。就是更新 map。数据库就更新数据库,同步 job 状态。
接着返回状态,回到了 TransactionTemplate#execute(TransactionCallback action)。提交事务,返回状态。
又回到TaskletStep#doExecute(StepExecution stepExecution)
中stepOperations.iterate(new StepContextRepeatCallback(stepExecution) {
里面。
这里面其实是在RepeatTemplate#iterate(RepeatCallback callback)
调用的executeInternal(callback)
这里。这里调用的具体逻辑下面仔细讲解。
先看TaskletStep#doExecute(StepExecution stepExecution)
中chunkListener.afterChunk(chunkContext);
。执行ChunkListener#afterChunk
。后面 interruptionPolicy.checkInterrupted(stepExecution); 判断线程是否被中短。
继续往下执行,回到 RepeatTemplate#executeInternal(final RepeatCallback callback)
。
executeAfterInterceptors(context, result);
执行RepeatListener#after
。
isComplete(context, result) || isMarkedComplete(context) || !deferred.isEmpty()
顺序在这里可能很重要。
protected boolean isComplete(RepeatContext context, RepeatStatus result) {
boolean complete = completionPolicy.isComplete(context, result);
if (complete) {
logger.debug("Repeat is complete according to policy and result value.");
}
return complete;
}
状态是否为RepeatStatus#CONTINUABLE
。
private boolean isMarkedComplete(RepeatContext context) {
boolean complete = context.isCompleteOnly();
if (context.getParent() != null) {
complete = complete || isMarkedComplete(context.getParent());
}
if (complete) {
logger.debug("Repeat is complete according to context alone.");
}
return complete;
}
检查上一层。
deferred.isEmpty()
在 doHandle(throwable, context, deferred);
中的,如果有异常,这个里面就有值。
现在没有进入,running
还是 true,所以继续执行。这个RepeatTemplate
是外层的,里面真正执行ItemReader
和ItemWriter
的是一个叫RepeatOperations chunkOperations
的RepeatTemplate
执行的。
执行完了之后,就到了result = result.and(waitForResults(state));
。为 ture,这个值在这个地方没有用的。
剩下就是RepeatListener#close
。context.close();
回到 AbstractStep#execute(StepExecution stepExecution)
。剩下的都是一些资源清除,catch 块里面的内容了,感兴趣的可以自己看看。
关键的就是
getJobRepository().updateExecutionContext(stepExecution);
getJobRepository().update(stepExecution);
close(stepExecution.getExecutionContext());
doExecutionRelease();
回到SimpleStepHandler#handleStep(Step step, JobExecution execution)
。刚刚我们所有方法都是在step.execute(currentStepExecution);
中执行的。继续也是一些异常处理,关键就下面几句。
currentStepExecution.getExecutionContext().put("batch.executed", true);
jobRepository.updateExecutionContext(execution);
回到SimpleJob#doExecute(JobExecution execution)
里面,我们把所有的steps
执行完了。然后状态更改。
execution.upgradeStatus(stepExecution.getStatus());
execution.setExitStatus(stepExecution.getExitStatus());
回到AbstractJob#execute(JobExecution execution)
,里面的doExecute(execution);
执行完成,后面又是一些状态更新,数据保存。
listener.afterJob(execution);
:JobExecutionListener#afterJob
jobRepository.update(execution);
JobSynchronizationManager.release();
至此就基本看完了里面大体流程。一些细节方面在下面类里面主要讲解。
BatchAutoConfiguration
BatchConfigurerConfiguration
DefaultBatchConfigurer
首先Autowired
注入DataSource
。这里就要区分是手动创建还是注入 spring 中的了。手动创建需要手动执行initialize()
方法,spring
创建的因为加了@PostConstruct
注解。会自动执行。
DefaultBatchConfigurer#initialize()
try {
if (dataSource == null) {
logger.warn("No datasource was provided...using a Map based JobRepository");
if (getTransactionManager() == null) {
logger.warn("No transaction manager was provided, using a ResourcelessTransactionManager");
this.transactionManager = new ResourcelessTransactionManager();
}
MapJobRepositoryFactoryBean jobRepositoryFactory = new MapJobRepositoryFactoryBean(getTransactionManager());
jobRepositoryFactory.afterPropertiesSet();
this.jobRepository = jobRepositoryFactory.getObject();
MapJobExplorerFactoryBean jobExplorerFactory = new MapJobExplorerFactoryBean(jobRepositoryFactory);
jobExplorerFactory.afterPropertiesSet();
this.jobExplorer = jobExplorerFactory.getObject();
} else {
this.jobRepository = createJobRepository();
this.jobExplorer = createJobExplorer();
}
this.jobLauncher = createJobLauncher();
} catch (Exception e) {
throw new BatchConfigurationException(e);
}
很简单的代码,dataSource
不存在,jobExplorer
和jobRepository
就使用MapJobRepositoryFactoryBean
和MapJobExplorerFactoryBean
这里还有个private PlatformTransactionManager transactionManager
,如果 spring 管理的DefaultBatchConfigurer
类,使用DataSourceTransactionManager
,否则使用ResourcelessTransactionManager
。这俩个类一个是 spring 事务相关的不做多讲解
Listener
里面涉及到一大堆的监听器,处理器。每个类基本都有讲到。 类上的注释非常重要~!
首先校验,然后使用stepExecution
保存数据。ecDao 保存 ExecutionContext。