Python: improving time required to load 1.5G CSV file
In previous post we discussed how we can search for specific item in 1.5G CSV file with 132M records. And we improved search from 73625.309 ms (more than a minute) to just ~0.005 ms - almost 15 million times faster. Which is pretty impressive improvement as per my understanding.
But there is still one bottle neck that can be improved - time required for first initial scan of the file. Let's try to improve this in this article.
All source files can be found in JFF-Bohdan/item_lookup