您好,登錄后才能下訂單哦!
接著上一篇,將mysql的數(shù)據(jù)導(dǎo)入kafka中
public static void main(String[] arg) throws Exception {
TypeInformation[] fieldTypes = new TypeInformation[] { BasicTypeInfo.STRING_TYPE_INFO, BasicTypeInfo.STRING_TYPE_INFO };
String[] fieldNames = new String[] { "name", "address" };
RowTypeInfo rowTypeInfo = new RowTypeInfo(fieldTypes, fieldNames);
JDBCInputFormat jdbcInputFormat = JDBCInputFormat.buildJDBCInputFormat().setDrivername("com.mysql.jdbc.Driver")
.setDBUrl("jdbc:mysql://ip:3306/tablespace?characterEncoding=utf8")
.setUsername("user").setPassword("root")
.setQuery("select LOGIC_CODE, SHARE_LOG_CODE from table").setRowTypeInfo(rowTypeInfo).finish();
final ExecutionEnvironment env = ExecutionEnvironment.getExecutionEnvironment();
DataSource<Row> s = env.createInput(jdbcInputFormat);
BatchTableEnvironment tableEnv = new BatchTableEnvironment(env, TableConfig.DEFAULT());
tableEnv.registerDataSet("t2", s);
Table tapiResult = tableEnv.scan("t2");
System.out.println("schema is:");
tapiResult.printSchema();
Table query = tableEnv.sqlQuery("select name, address from t2");
DataSet<Result> ds= tableEnv.toDataSet(query, Result.class);
DataSet<String> temp=ds.map(new MapFunction<Result, String>() {
@Override
public String map(Result result) throws Exception {
String name = result.name;
String value = result.address;
return name+":->:"+value;
}
});
logger.info("read db end");
KafkaOutputFormat kafkaOutput = KafkaOutputFormat.buildKafkaOutputFormat()
.setBootstrapServers("ip:9092").setTopic("search_test_whk").setAcks("all").setBatchSize("1000")
.setBufferMemory("100000").setLingerMs("1").setRetries("2").finish();
temp.output(kafkaOutput);
logger.info("write kafka end");
env.execute("Flink add data source");
}
免責(zé)聲明:本站發(fā)布的內(nèi)容(圖片、視頻和文字)以原創(chuàng)、轉(zhuǎn)載和分享為主,文章觀點(diǎn)不代表本網(wǎng)站立場(chǎng),如果涉及侵權(quán)請(qǐng)聯(lián)系站長(zhǎng)郵箱:is@yisu.com進(jìn)行舉報(bào),并提供相關(guān)證據(jù),一經(jīng)查實(shí),將立刻刪除涉嫌侵權(quán)內(nèi)容。