jackg Posted February 18, 2008 Share Posted February 18, 2008 I need to sort and de-duplicate terms in a very large file. Too big for array sort. I can't load into an array a file of, say, 50,000 to 80,000 terms. How can I approach doing this? Thanks, Jackg Louisiana Quote Link to comment Share on other sites More sharing options...
effigy Posted February 18, 2008 Share Posted February 18, 2008 Have you tried Unix's sort -u? Quote Link to comment Share on other sites More sharing options...
jackg Posted February 18, 2008 Author Share Posted February 18, 2008 Yes, sort -u worked on unix, now I have to re-write all my code for a windows box? Dam. I can't get access to windows commands -- not that I want them! Thanks, Jackg Quote Link to comment Share on other sites More sharing options...
effigy Posted February 18, 2008 Share Posted February 18, 2008 Perl and File::Sort perhaps? If not, there are Unix utilities available for Windows: Cygwin. Quote Link to comment Share on other sites More sharing options...
Recommended Posts
Join the conversation
You can post now and register later. If you have an account, sign in now to post with your account.