I have written a small proof of concept Spring Boot / Batch application that uses JDBC to page through a very large database table (potentially billions of rows) and converts the rows read to XML before writing them out to an S3 bucket. Currently my PoC has a single job that commences on start up and will keep running until the table has been completely consumed. It runs against test data, which is only a few hundred rows. For my single job I am using:
JdbcPagingItemReader
- A custom
ItemProcessor
to convert rows to XML
- A custom
ItemWriter
to write XML to S3
I have a number of problems which I don't know how to address using Spring Batch. Having read the manual, nothing is jumping out at me as a solution for how I might progress. For the sake of argument lets say the table contains 1 billion rows of Person information and looks like this:
与恶龙缠斗过久,自身亦成为恶龙;凝视深渊过久,深渊将回以凝视…