下面是我的剧本:
import glob,os, csv
from sqlalchemy import *
count = 0
served_imsi = []
served_imei = []
served_msisdn = []
location_area_code = []
routing_area = []
cell_identity = []
service_area_code = []
s_charging_characteristics = []
plmn_id = []
path = '/home/cneps/cdr/*.cdr'
for file in glob.glob(path):
f = open(file)
for lines in f:
served_imsi.append(lines[17:17+16])
served_imei.append(lines[47:47+16])
served_msisdn.append(lines[65:65+18])
sgsn_address.append(lines[83:83+32])
ggsn_address.append(lines[115:115+32])
charging_id.append(lines[147:147+10])
apn_network.append(lines[157:157+63])
location_area_code.append(lines[296:296+4])
routing_area.append(lines[300:300+2])
cell_identity.append(lines[302:302+4])
service_area_code.append(lines[306:306+4])
s_charging_characteristics.append(lines[325:325+2])
plmn_id.append(lines[327:327+6])
db = create_engine('sqlite:///TIM_CDR.db',echo=True)
metadata = MetaData(db)
CDR1 = Table('CDR1', metadata, autoload=False)
i = CDR1.insert()
while count < len(served_imei):
i.execute(Served_IMSI=served_imsi[count], Served_IMEI=served_imei[count], Served_MSISDN=served_msisdn[count], SGSN_Address=sgsn_address[count], GGSN_Address=ggsn_address[count], Charging_ID=charging_id[count], APN_Network=apn_network[count], LAC=location_area_code[count], RAC=routing_area[count], Cell_Identity=cell_identity[count], Service_Area_Code=service_area_code[count], S_Charging_Characteristics=s_charging_characteristics[count], PLMN_ID=plmn_id[count])
count += 1
因为我要插入数据库的这些数据像是10万行,所以需要花很多时间才能完成。在
要花30分钟才能完成。在
我已经读过了,我知道我可能应该使用事务处理,但我不知道怎么做。在
有人能在我的代码里给我举个例子吗?如何在所有事情之后使用事务提交?在
那太好了,谢谢。在
请参见这个答案中的
test_alchemy_core
函数:Why is SQLAlchemy insert with sqlite 25 times slower than using sqlite3 directly?它向您展示了如何在一个批处理中执行多个插入。你的问题是,你一个接一个地执行插入,这总是非常缓慢的。在相关问题 更多 >
编程相关推荐