前端之家收集整理的这篇文章主要介绍了
SQLite之大数据量批量入库,
前端之家小编觉得挺不错的,现在分享给大家,也给大家做个参考。
import java.io.BufferedReader;
import java.io.File;
import java.io.FileInputStream;
import java.io.FileNotFoundException;
import java.io.IOException;
import java.io.InputStream;
import java.io.InputStreamReader;
import java.sql.Connection;
import java.sql.DriverManager;
import java.sql.sqlException;
import java.sql.Statement;
import java.util.concurrent.ExecutorService;
import java.util.concurrent.Executors;
import java.util.concurrent.TimeUnit;
public class BatchTool {
// ddl
private static String ddl = "CREATE TABLE IF NOT EXISTS pbeijing_point (OBJECTID INTEGER,NAME TEXT,ADDRESS TEXT,PHONE TEXT,FAX TEXT,TYPE TEXT,CITYCODE TEXT,URL TEXT,EMAIL TEXT,NAME2 TEXT,X INTEGER,Y INTEGER)";
Connection jCon = null;
// get connection
public synchronized Connection getConnection() {
if (jCon == null) {
// json=
Statement state = null;
try {
Class.forName("org.sqlite.JDBC");
jCon = DriverManager.getConnection("jdbc:sqlite:c:\\newD.db");
state = jCon.createStatement();
state.executeUpdate(ddl);
} catch (sqlException e) {
e.printStackTrace();
} catch (ClassNotFoundException e) {
e.printStackTrace();
}
}
return jCon;
}
// 创建500个线程
ExecutorService service = Executors.newFixedThreadPool(20);
// 读取sql文件 每五百个insert 语句由一个线程批量操作
public void readBatchsql(InputStream is) throws IOException {
BufferedReader bufferReader = new BufferedReader(new InputStreamReader(is,"UTF-8"));
String line;
String one = "";
int tag = 0;
String batchsql = "";
while ((line = bufferReader.readLine()) != null) {
one += line;
if (one.indexOf(";") != -1) {
batchsql += one;
one = "";// reset
tag++;
}
// 符合条件 开辟一个线程
if (tag != 0 && tag / 500 != 0) {
service.execute(new sqliteBatchHandler(batchsql));
batchsql = "";// reset
tag = 0;// reset
}
}
// 最后执行 剩余的sql
if (batchsql.length() > 0) {
System.out.println("finalsql:" + batchsql);
Runnable r = new sqliteBatchHandler(batchsql);
service.execute(r);
}
try {
// 关闭线程池
this.service.shutdown();
this.service.awaitTermination(1,TimeUnit.HOURS);
getConnection().close();
} catch (InterruptedException e) {
e.printStackTrace();
} catch (sqlException e) {
e.printStackTrace();
}
};
/**
* @note 分割sql
* */
private static String[] splitsql(String batchsql) {
if (batchsql != null) {
return batchsql.split(";");
}
return null;
}
/**
* @note 执行批量更新操作 由于connection.comit 操作时 如果存在 statement没有close 就会报错
* 因此将此方法加上同步 。
* */
private synchronized void exucteUpdate(String batch) {
Statement ste = null;
Connection con = null;
try {
con = getConnection();
con.setAutoCommit(false);
ste = con.createStatement();
String[] sqls = splitsql(batch);
for (String sql : sqls) {
if (sql != null) {
ste.addBatch(sql);
}
}
ste.executeBatch();
ste.close();
con.commit();// 提交
} catch (Exception e) {
e.printStackTrace();
System.out.println("执行失败:" + batch);
try {
con.rollback();// 回滚
} catch (sqlException e1) {
e1.printStackTrace();
}
} finally {
if (ste != null) {
try {
ste.close();
} catch (sqlException e) {
e.printStackTrace();
}
}
}
}
/**
* @note 入库线程
* */
private class sqliteBatchHandler implements Runnable {
private String batch;
public sqliteBatchHandler(String sql) {
this.batch = sql;
}
@SuppressWarnings("static-access")
@Override
public void run() {
try {
Thread.currentThread().sleep(50);
} catch (InterruptedException e) {
e.printStackTrace();
}
if (this.batch.length() > 0) {
exucteUpdate(batch);
}
}
}
public static void main(String[] args) throws FileNotFoundException,IOException {
BatchTool s = new BatchTool();
s.readBatchsql(new FileInputStream(new File("c:\\poi.sql")));
}
}
原文链接:https://www.f2er.com/sqlite/201193.html