Eloquent chunk() missing half the results

Quick answer: Use chunkById() instead of chunk().

When updating or deleting records while iterating over them, any changes to the primary key or foreign keys could affect the chunk query. This could potentially result in records not being included in the results.

The explanation can be found in the Laravel documentation:

Here is the solution example:

DB::table('users')->where('active', false)
    ->chunkById(100, function ($users) {
        foreach ($users as $user) {
            DB::table('users')
                ->where('id', $user->id)
                ->update(['active' => true]);
        }
    });

If you are updating database records while chunking results, your chunk results could change in unexpected ways. If you plan to update the retrieved records while chunking, it is always best to use the chunkById method instead. This method will automatically paginate the results based on the record's primary key.

(end of the update)

The original answer:

I had the same problem - only half of the total results were passed to the callback function of the chunk() method.

Here is the code which had problems:

Transaction::whereNull('processed')->chunk(100, function ($transactions) {
    $transactions->each(function($transaction){
        $transaction->process();
    });
});

I used Laravel 5.4 and managed to solve the problem replacing the chunk() method with cursor() method and changing the code accordingly:

foreach (Transaction::whereNull('processed')->cursor() as $transaction) {
    $transaction->process();
}

Even though the answer doesn't address the problem itself, it provides a valuable solution.


Imagine you are using chunk method to delete all of the records. The table has 2,000,000 records and you are going to delete all of them by 1000 chunks.

$query->orderBy('id')->chunk(1000, function ($items) {
    foreach($items as $item) {
        $item->delete();
    }
});

It will delete the first 1000 records by getting first 1000 records in a query like this:

SELECT * FROM table ORDER BY id LIMIT 0,1000

And then the other query from chunk method is:

SELECT * FROM table ORDER BY id LIMIT 1000,2000

Our problem is here, that we delete 1000 records and then getting results from 1000 to 2000. Actually we are missing first 1000 records and this means that we are not going to delete 1000 records in first step of chunk! This scenario will be the same for other steps. In each step we are going to miss 1000 records and this is the reason that we are not getting best result in these situations.

I made an example for deletion because this way we could know the exact behavior of chunk method.


UPDATE:

You can use chunkById() for deleting safely.

Read more here:

http://laravel.at.jeffsbox.eu/laravel-5-eloquent-builder-chunk-chunkbyid https://laravel.com/api/5.4/Illuminate/Database/Eloquent/Builder.html#method_chunkById


For anyone looking for a bit of code that solves this, here you go:

while (Model::where('x', '>', 'y')->count() > 0)
{
    Model::where('x', '>', 'y')->chunk(10, function ($models)
    {
        foreach ($models as $model)
        {
            $model->delete();
        }
    });
}

The problem is in the deletion / removal of the model while chunking away at the total. Including it in a while loop makes sure you get them all! This example works when deleting Models, change the while condition to suit your needs!