Hi everyone!
PHP question (new to PHP after 10 years of Perl, arrggghhh!).
I have a 100mb file that I want to send to another server.
I have managed to read the file in and "post" it without curl (cannot use curl for this app). Everything is working fine on smaller files.
However, with the larger files, PHP complains about not being able to allocate memory.
Is there a way to open a file, line by line, and send it as a post ALSO line by line?
This way nothing is held in ram thus preventing my errors and getting around strict limitations.
Chris
Here's my current code that errors with large files:
<?php
ini_set('display_errors',1);
error_reporting(E_ALL|E_STRICT);
$resp = do_post_request("/local/file.txt","http://www.mysite.com/receivedata.php");
exit;
function do_post_request($file,$url){
$fileHandle = fopen($file, "rb");
$fileContents = stream_get_contents($fileHandle);
fclose($fileHandle);
$params = array(
'http' => array
(
'method' => 'POST',
'header'=>"Content-Type: multipart/form-data\r\n",
'content' => $fileContents
)
);
$ctx = stream_context_create($params);
$fp = fopen($url, 'rb', false, $ctx);
$response = stream_get_contents($fp);
return $response;
}
?>