@Test
public void testExtractor() throws Exception {
// setup
Schema schema = new Schema("testExtractor");
schema.addColumn(new Text("TextCol"));
ExtractorContext context = new ExtractorContext(null, writerMock, schema);
LinkConfiguration linkConfig = new LinkConfiguration();
FromJobConfiguration jobConfig = new FromJobConfiguration();
KiteDatasetPartition partition = new KiteDatasetPartition();
partition.setUri("dataset:hdfs:/path/to/dataset");
OngoingStubbing<Object[]> readRecordMethodStub = when(executorMock.readRecord());
final int NUMBER_OF_ROWS = 1000;
for (int i = 0; i < NUMBER_OF_ROWS; i++) {
// TODO: SQOOP-1616 will cover more column data types
readRecordMethodStub = readRecordMethodStub.thenReturn(new Object[]{});
}
readRecordMethodStub.thenReturn(null);
// exercise
extractor.extract(context, linkConfig, jobConfig, partition);
// verify
verify(writerMock, times(NUMBER_OF_ROWS)).writeArrayRecord(
any(Object[].class));
}
TestKiteExtractor.java 文件源码
java
阅读 26
收藏 0
点赞 0
评论 0
项目:sqoop-on-spark
作者:
评论列表
文章目录