Avatar of blink10
 asked on

Linux box killing my script...

I trying to run a while loop to process my product's records.

However....I have about 8 million records and every time I try to run it, it just does nothing for 3 minutes and then just says killed.

Why does this while loop kill it? Below I have the code will is causing the problems....ideas on how to modify it? (when i limit it to 10 it runs ok)

I have a feeling it a memory issue, by why cant it process this number of records?

$resultp = mysql_query("SELECT * FROM LinkshareProducts WHERE P_Id<>''");

while($row = mysql_fetch_array($resultp)) {


//special scripting to get isbn number

$isbnLENGTN=strlen($isbn); // should be 10 or 13 long

$Name = mysql_real_escape_string($Name);

$pn = mysql_real_escape_string($mid);
$resultp1 = mysql_query("SELECT * FROM Stores WHERE MID='$pn'");
$row1 = mysql_fetch_array($resultp1);



$upcLENGTH=strlen($UPC); // should be 12 long




$Description = mysql_real_escape_string($Description);
$a1 = mysql_real_escape_string($a1);
$a2 = mysql_real_escape_string($a2);
$a3 = mysql_real_escape_string($a3);
$a4 = mysql_real_escape_string($a4);
$a5 = mysql_real_escape_string($a5);
$a6 = mysql_real_escape_string($a6);
$a7 = mysql_real_escape_string($a7);

$InStock = mysql_real_escape_string($InStock);

$t1 = mysql_real_escape_string($t1);
$t2 = mysql_real_escape_string($t2);

if ($Instock!="no"){
$Description = mysql_real_escape_string($Description);
$Brand = mysql_real_escape_string($Brand);
$Image = mysql_real_escape_string($Image);
$NewUsed = mysql_real_escape_string($NewUsed);
$manid = mysql_real_escape_string($manid);

mysql_query("INSERT INTO products (id, name, description, brand, image, NewUsed, tier1, tier2, source, a1, a2, a3, a4, a5, a6, a7, network) VALUES ('$id', '$Name', '$Description', '$Brand', '$Image', '$NewUsed', '$t1', '$t2', '$source', '$a1', '$a2', '$a3', '$a4', '$a5', '$a6', '$a7', '$network')")or die(mysql_error());

Open in new window


Avatar of undefined
Last Comment
Beverley Portlock

8/22/2022 - Mon
Dave Baldwin

There are always memory and time limits.  If your rows have only 100 characters, that's 800MegaBytes of data.

Why don't you break it up into smaller batches like 100,000 rows?  Add an 'ORDER BY' clause and keep track of the last value in each batch and use that to start the next batch.

can u show me how i would break it up, i am not getting exactly what you are saying
Dave Baldwin

The first is similar to what you've done.  The rest of the time, you take the last P_Id from the previous run and get 100000 items that are greater than it.  This does assume that P_Id is unique so that the ORDER BY sort works properly.  Note that sorting an alphanumeric column automatically puts '' at the beginning.
// first time
$resultp = mysql_query("SELECT * FROM LinkshareProducts WHERE P_Id<>'' ORDER BY P_Id LIMIT 100000");
// the rest of the times
$resultp = mysql_query("SELECT * FROM LinkshareProducts WHERE P_Id > 'thelastP_Id' ORDER BY P_Id LIMIT 100000");

Open in new window

I started with Experts Exchange in 2004 and it's been a mainstay of my professional computing life since. It helped me launch a career as a programmer / Oracle data analyst
William Peck

dumb question but how do i get the last pid and make the rest of the times while loop to keep running as results are available?
Dave Baldwin

Log in or sign up to see answer
Become an EE member today7-DAY FREE TRIAL
Members can start a 7-Day Free trial then enjoy unlimited access to the platform
Sign up - Free for 7 days
Learn why we charge membership fees
We get it - no one likes a content blocker. Take one extra minute and find out why we block content.
Not exactly the question you had in mind?
Sign up for an EE membership and get your own personalized solution. With an EE membership, you can ask unlimited troubleshooting, research, or opinion questions.
ask a question
Beverley Portlock

CHeck that table 'stores' has an index defined on column 'MID'. If not then add one and you will see a vast improvement in performance, but DaveBaldwin's comment about repeatedly re-running SELECTs is still relevant even if there is an index problem.