I use huge data files, sometimes I only need to know the number of lines in these files, usually I open them up and read them line by line until I reach the end of the file
I was wondering if there is a smarter way to do that
...
I am working on a C++ program and the compiled object code from a single 1200-line file (which initializes a rather complex state machine) comes out to nearly a megabyte. What could be making the file so large? Is there a way I can find what takes space inside the object file?
...
When is PostedFile.InputStream available when uploading a large file?
I'd like to pass a Stream to another process and I'm hoping that if a large file was being uploaded that I can pass the Stream straight to that new process w/o writing to the file system. Since the process and/or upload could take a while, I'm wondering if I can star...
I need to keep under version some large files (some Gigs).
I don't need, and I can't keep under version all the version of the files.
I want to be able to remove from my VCS large files version at some moment.
What control version system could I use?
EDIT:
The files that I want to keep under version control are big .zip files or ISO...
Hello,
I am trying to write a Java class to extract a large zip file containing ~74000 XML files. I get the following exception when attempting to unzip it utilizing the java zip library:
java.util.zip.ZipException: too many entries in ZIP file
Unfortunately due to requirements of the project I can not get the zip broken down before...
Hi!
I want to upload a large file of size 10M max to my mysql database. Using .htaccess i changed the PHP's own file upload limit to "10485760" = 10M, i am able to upload files upto 10M size without any problem.
But i can not insert the file in database if it is more that 1M in size.
i am using file_get_contents to read all file data ...
Hi!
I want to upload a large file of size 10M max to my mysql database. Using .htaccess i changed the PHP's own file upload limit to "10485760" = 10M, i am able to upload files upto 10M size without any problem.
But i can not insert the file in database if it is more that 1M in size.
i am using file_get_contents to read all file data ...
I have 1 source server that contains about 30GB of files that I want to copy to 7 other servers. I currently SCP the files over to the first four servers and when that transfer is complete SCP to the last 3 servers.
Would using BitTorrent be faster to distribute the files to all 7 at once? The connection internally is already fast and...
I have a very large CSV file and I need to write an app that will parse it but using the >6GB file to test against is painful, is there a simple way to extract the first hundred or two lines without having to load the entire file into memory?
The file resides on a Debian server.
...
I have an ASP.Net MVC application that allows users to upload images. When I try to upload a really large file (400MB) I get an error.
I assumed that my image processing code (home brew) was very inefficient, so I decided I would try using a third party library to handle the image processing parts.
Because I'm using TDD, I wanted to fi...
I have some really large files for example 320 MB tif file with 14000 X 9000 pixels.
The operations I need to perform are basically scaling the images to get smaller versions of it and breaking the image into tiles.
My code works fine with small files and I use the .Net Bitmap objects but I will occasionally get Out of Memory exception...
Hi there. I am looking for opinions of how to handle large binary files on which my source code (web application) is dependent. We are currently discussing several alternatives:
Copy the binary files by hand.
Pro: Not sure.
Contra: I am strongly against this, as it increases the likelihood of errors when setting up a new site/migratin...
Hi,
I have a php form that has a bunch of checkboxes that all contain links to files. Once a user clicks on which checkboxes (files) they want, it then zips up the files and forces a download.
I got a simple php zip force download to work, but when one of the files is huge or if someone lets say selects the whole list to zip up and ...
Hello,
I have a big file with thousands of lines of thousands of characters.
I move the cursor to 3000th character.
If I use PageDown or <CTRL>-D, the file will scroll but the cursor will come back to the first no-space character.
There's is an option to set to keep the cursor in the same column after a such scroll ?
I have the beh...
I need to perform a simple grep and other manipulations on large files in Java. I am not that familiar with the Java NIO utilities, but I am assuming that is what I need to use. What resources or helpful tips do you have for reading/writing large files. Also, I am working on a SWT application and need to display parts of that data wit...
Hi,
I am using expat parser to parse an XML file of around 15 GB . The problem is it throws an "Out of Memory" error and the program aborts .
I want to know has any body faced a similar issue with the expat parser or is it a known bug and has been rectified in later versions ?
...
I have to import a large mysql dump (up to 10G). However the sql dump already predefined with a database structure with index definition. I want to speed up the db insert by removing the index and table definition.
That means I have to remove/edit the first few lines of a 10G text file. What is the most efficient way to do this on li...
I am struggling to port a Perl program to Java, and learning Java as I go. A central component of the original program is a Perl module that does string prefix lookups in a +500 GB sorted text file using binary search
(essentially, "seek" to a byte offset in the middle of the file, backtrack to nearest newline, compare line prefix with ...
I have a 384MB text file with 50 million lines. Each line contains 2 space-separated integers: a key and a value. The file is sorted by key. I need an efficient way of looking up the values of a list of about 200 keys in Python.
My current approach is included below. It takes 30 seconds. There must be more efficient Python foo to get th...
Given:
file a.txt containing many millions
of lines (say, one sentence per line) (2.6 gb!)
file b.txt containing 830k lines with pairs "[word1] [word2]"
Question:
how to perform the most efficient replacement of each word1 by word2 for every of 830k tuples (w1, w2) in the huge text file?
Naive methods like sed, perl, python etc. wo...