As of now I use the following python code:
file = open(filePath, "r")
lines=file.readlines()
file.close()
Say my file has several lines (10,000 or more), then my program becomes slow if I do this for more than one file. Is there a way to speed this up in Python? Reading various links I understand that readlines stores the lines of file in memory thats why the code gets slow.
I have tried the following code as well and the time gain I got is 17%.
lines=[line for line in open(filePath,"r")]
Is there any other module in python2.4 (which I might have missed). Thanks, Sandhya
This gives you an iterator that reads the file object one line at a time and then discards the previous line from memory.
Short answer: don't assign the lines to a variable, just perform whatever operations you need inside the loop.