Compare commits
4 Commits
a8dafb23c5
...
1b3e0ed30d
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
1b3e0ed30d | ||
|
|
0b607f86cb | ||
|
|
30662b97d5 | ||
|
|
cbfe2c4aa9 |
@@ -20,10 +20,12 @@
|
|||||||
private const MAX_EXECUTION_TIME = 86400; // 24 hours in seconds
|
private const MAX_EXECUTION_TIME = 86400; // 24 hours in seconds
|
||||||
private const FILENAME = 'ST.AA.ARRANGEMENT.csv';
|
private const FILENAME = 'ST.AA.ARRANGEMENT.csv';
|
||||||
private const DISK_NAME = 'sftpStatement';
|
private const DISK_NAME = 'sftpStatement';
|
||||||
|
private const CHUNK_SIZE = 1000; // Process data in chunks to reduce memory usage
|
||||||
|
|
||||||
private string $period = '';
|
private string $period = '';
|
||||||
private int $processedCount = 0;
|
private int $processedCount = 0;
|
||||||
private int $errorCount = 0;
|
private int $errorCount = 0;
|
||||||
|
private array $arrangementBatch = [];
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Create a new job instance.
|
* Create a new job instance.
|
||||||
@@ -61,6 +63,7 @@
|
|||||||
set_time_limit(self::MAX_EXECUTION_TIME);
|
set_time_limit(self::MAX_EXECUTION_TIME);
|
||||||
$this->processedCount = 0;
|
$this->processedCount = 0;
|
||||||
$this->errorCount = 0;
|
$this->errorCount = 0;
|
||||||
|
$this->arrangementBatch = [];
|
||||||
}
|
}
|
||||||
|
|
||||||
private function processPeriod()
|
private function processPeriod()
|
||||||
@@ -111,10 +114,23 @@
|
|||||||
|
|
||||||
$headers = (new TempArrangement())->getFillable();
|
$headers = (new TempArrangement())->getFillable();
|
||||||
$rowCount = 0;
|
$rowCount = 0;
|
||||||
|
$chunkCount = 0;
|
||||||
|
|
||||||
while (($row = fgetcsv($handle, 0, self::CSV_DELIMITER)) !== false) {
|
while (($row = fgetcsv($handle, 0, self::CSV_DELIMITER)) !== false) {
|
||||||
$rowCount++;
|
$rowCount++;
|
||||||
$this->processRow($row, $headers, $rowCount, $filePath);
|
$this->processRow($row, $headers, $rowCount, $filePath);
|
||||||
|
|
||||||
|
// Process in chunks to avoid memory issues
|
||||||
|
if (count($this->arrangementBatch) >= self::CHUNK_SIZE) {
|
||||||
|
$this->saveBatch();
|
||||||
|
$chunkCount++;
|
||||||
|
Log::info("Processed chunk $chunkCount ({$this->processedCount} records so far)");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Process any remaining records
|
||||||
|
if (!empty($this->arrangementBatch)) {
|
||||||
|
$this->saveBatch();
|
||||||
}
|
}
|
||||||
|
|
||||||
fclose($handle);
|
fclose($handle);
|
||||||
@@ -127,22 +143,29 @@
|
|||||||
if (count($headers) !== count($row)) {
|
if (count($headers) !== count($row)) {
|
||||||
Log::warning("Row $rowCount in $filePath has incorrect column count. Expected: " .
|
Log::warning("Row $rowCount in $filePath has incorrect column count. Expected: " .
|
||||||
count($headers) . ", Got: " . count($row));
|
count($headers) . ", Got: " . count($row));
|
||||||
|
$this->errorCount++;
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
$data = array_combine($headers, $row);
|
$data = array_combine($headers, $row);
|
||||||
$this->saveRecord($data, $rowCount, $filePath);
|
$this->addToBatch($data, $rowCount, $filePath);
|
||||||
}
|
}
|
||||||
|
|
||||||
private function saveRecord(array $data, int $rowCount, string $filePath)
|
/**
|
||||||
|
* Add record to batch instead of saving immediately
|
||||||
|
*/
|
||||||
|
private function addToBatch(array $data, int $rowCount, string $filePath)
|
||||||
: void
|
: void
|
||||||
{
|
{
|
||||||
try {
|
try {
|
||||||
if ($data['arrangement_id'] !== 'arrangement_id') {
|
if ($data['arrangement_id'] !== 'arrangement_id') {
|
||||||
TempArrangement::updateOrCreate(
|
// Add timestamp fields
|
||||||
['arrangement_id' => $data['arrangement_id']], // key to find existing record
|
$now = now();
|
||||||
$data // data to update or create
|
$data['created_at'] = $now;
|
||||||
);
|
$data['updated_at'] = $now;
|
||||||
|
|
||||||
|
// Add to batch
|
||||||
|
$this->arrangementBatch[] = $data;
|
||||||
$this->processedCount++;
|
$this->processedCount++;
|
||||||
}
|
}
|
||||||
} catch (Exception $e) {
|
} catch (Exception $e) {
|
||||||
@@ -151,6 +174,32 @@
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Save batched records to the database
|
||||||
|
*/
|
||||||
|
private function saveBatch()
|
||||||
|
: void
|
||||||
|
{
|
||||||
|
try {
|
||||||
|
if (!empty($this->arrangementBatch)) {
|
||||||
|
// Bulk insert/update arrangements
|
||||||
|
TempArrangement::upsert(
|
||||||
|
$this->arrangementBatch,
|
||||||
|
['arrangement_id'], // Unique key
|
||||||
|
array_diff((new TempArrangement())->getFillable(), ['arrangement_id']) // Update columns
|
||||||
|
);
|
||||||
|
|
||||||
|
// Reset batch after processing
|
||||||
|
$this->arrangementBatch = [];
|
||||||
|
}
|
||||||
|
} catch (Exception $e) {
|
||||||
|
Log::error("Error in saveBatch: " . $e->getMessage());
|
||||||
|
$this->errorCount += count($this->arrangementBatch);
|
||||||
|
// Reset batch even if there's an error to prevent reprocessing the same failed records
|
||||||
|
$this->arrangementBatch = [];
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
private function cleanup(string $tempFilePath)
|
private function cleanup(string $tempFilePath)
|
||||||
: void
|
: void
|
||||||
{
|
{
|
||||||
|
|||||||
@@ -20,6 +20,7 @@
|
|||||||
private const MAX_EXECUTION_TIME = 86400; // 24 hours in seconds
|
private const MAX_EXECUTION_TIME = 86400; // 24 hours in seconds
|
||||||
private const FILENAME = 'ST.ATM.TRANSACTION.csv';
|
private const FILENAME = 'ST.ATM.TRANSACTION.csv';
|
||||||
private const DISK_NAME = 'sftpStatement';
|
private const DISK_NAME = 'sftpStatement';
|
||||||
|
private const CHUNK_SIZE = 1000; // Process data in chunks to reduce memory usage
|
||||||
private const HEADER_MAP = [
|
private const HEADER_MAP = [
|
||||||
'id' => 'transaction_id',
|
'id' => 'transaction_id',
|
||||||
'card_acc_id' => 'card_acc_id',
|
'card_acc_id' => 'card_acc_id',
|
||||||
@@ -43,6 +44,7 @@
|
|||||||
private string $period = '';
|
private string $period = '';
|
||||||
private int $processedCount = 0;
|
private int $processedCount = 0;
|
||||||
private int $errorCount = 0;
|
private int $errorCount = 0;
|
||||||
|
private array $atmTransactionBatch = [];
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Create a new job instance.
|
* Create a new job instance.
|
||||||
@@ -80,6 +82,7 @@
|
|||||||
set_time_limit(self::MAX_EXECUTION_TIME);
|
set_time_limit(self::MAX_EXECUTION_TIME);
|
||||||
$this->processedCount = 0;
|
$this->processedCount = 0;
|
||||||
$this->errorCount = 0;
|
$this->errorCount = 0;
|
||||||
|
$this->atmTransactionBatch = [];
|
||||||
}
|
}
|
||||||
|
|
||||||
private function processPeriod()
|
private function processPeriod()
|
||||||
@@ -136,9 +139,23 @@
|
|||||||
}
|
}
|
||||||
|
|
||||||
$rowCount = 0;
|
$rowCount = 0;
|
||||||
|
$chunkCount = 0;
|
||||||
|
|
||||||
while (($row = fgetcsv($handle, 0, self::CSV_DELIMITER)) !== false) {
|
while (($row = fgetcsv($handle, 0, self::CSV_DELIMITER)) !== false) {
|
||||||
$rowCount++;
|
$rowCount++;
|
||||||
$this->processRow($headerRow, $row, $rowCount, $filePath);
|
$this->processRow($headerRow, $row, $rowCount, $filePath);
|
||||||
|
|
||||||
|
// Process in chunks to avoid memory issues
|
||||||
|
if (count($this->atmTransactionBatch) >= self::CHUNK_SIZE) {
|
||||||
|
$this->saveBatch();
|
||||||
|
$chunkCount++;
|
||||||
|
Log::info("Processed chunk $chunkCount ({$this->processedCount} records so far)");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Process any remaining records
|
||||||
|
if (!empty($this->atmTransactionBatch)) {
|
||||||
|
$this->saveBatch();
|
||||||
}
|
}
|
||||||
|
|
||||||
fclose($handle);
|
fclose($handle);
|
||||||
@@ -151,15 +168,16 @@
|
|||||||
if (count($headerRow) !== count($row)) {
|
if (count($headerRow) !== count($row)) {
|
||||||
Log::warning("Row $rowCount in $filePath has incorrect column count. Expected: " .
|
Log::warning("Row $rowCount in $filePath has incorrect column count. Expected: " .
|
||||||
count($headerRow) . ", Got: " . count($row));
|
count($headerRow) . ", Got: " . count($row));
|
||||||
|
$this->errorCount++;
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
// Combine the header row with the data row
|
// Combine the header row with the data row
|
||||||
$rawData = array_combine($headerRow, $row);
|
$rawData = array_combine($headerRow, $row);
|
||||||
$this->mapAndSaveRecord($rawData, $rowCount, $filePath);
|
$this->mapAndAddToBatch($rawData, $rowCount, $filePath);
|
||||||
}
|
}
|
||||||
|
|
||||||
private function mapAndSaveRecord(array $rawData, int $rowCount, string $filePath)
|
private function mapAndAddToBatch(array $rawData, int $rowCount, string $filePath)
|
||||||
: void
|
: void
|
||||||
{
|
{
|
||||||
// Map the raw data to our model fields
|
// Map the raw data to our model fields
|
||||||
@@ -173,24 +191,53 @@
|
|||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
$this->saveRecord($data, $rowCount, $filePath);
|
$this->addToBatch($data, $rowCount, $filePath);
|
||||||
}
|
}
|
||||||
|
|
||||||
private function saveRecord(array $data, int $rowCount, string $filePath)
|
/**
|
||||||
|
* Add record to batch instead of saving immediately
|
||||||
|
*/
|
||||||
|
private function addToBatch(array $data, int $rowCount, string $filePath)
|
||||||
: void
|
: void
|
||||||
{
|
{
|
||||||
try {
|
try {
|
||||||
|
// Add timestamp fields
|
||||||
|
$now = now();
|
||||||
|
$data['created_at'] = $now;
|
||||||
|
$data['updated_at'] = $now;
|
||||||
|
|
||||||
// Create or update the record
|
// Add to batch
|
||||||
AtmTransaction::updateOrCreate(
|
$this->atmTransactionBatch[] = $data;
|
||||||
['transaction_id' => $data['transaction_id']],
|
|
||||||
$data
|
|
||||||
);
|
|
||||||
|
|
||||||
$this->processedCount++;
|
$this->processedCount++;
|
||||||
} catch (Exception $e) {
|
} catch (Exception $e) {
|
||||||
$this->errorCount++;
|
$this->errorCount++;
|
||||||
Log::error("Error processing row $rowCount in $filePath: " . $e->getMessage());
|
Log::error("Error processing ATM Transaction at row $rowCount in $filePath: " . $e->getMessage());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Save batched records to the database
|
||||||
|
*/
|
||||||
|
private function saveBatch()
|
||||||
|
: void
|
||||||
|
{
|
||||||
|
try {
|
||||||
|
if (!empty($this->atmTransactionBatch)) {
|
||||||
|
// Bulk insert/update ATM transactions
|
||||||
|
AtmTransaction::upsert(
|
||||||
|
$this->atmTransactionBatch,
|
||||||
|
['transaction_id'], // Unique key
|
||||||
|
array_values(array_diff(self::HEADER_MAP, ['id'])) // Update columns (all except transaction_id)
|
||||||
|
);
|
||||||
|
|
||||||
|
// Reset batch after processing
|
||||||
|
$this->atmTransactionBatch = [];
|
||||||
|
}
|
||||||
|
} catch (Exception $e) {
|
||||||
|
Log::error("Error in saveBatch: " . $e->getMessage());
|
||||||
|
$this->errorCount += count($this->atmTransactionBatch);
|
||||||
|
// Reset batch even if there's an error to prevent reprocessing the same failed records
|
||||||
|
$this->atmTransactionBatch = [];
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -20,10 +20,12 @@
|
|||||||
private const MAX_EXECUTION_TIME = 86400; // 24 hours in seconds
|
private const MAX_EXECUTION_TIME = 86400; // 24 hours in seconds
|
||||||
private const FILENAME = 'ST.AA.BILL.DETAILS.csv';
|
private const FILENAME = 'ST.AA.BILL.DETAILS.csv';
|
||||||
private const DISK_NAME = 'sftpStatement';
|
private const DISK_NAME = 'sftpStatement';
|
||||||
|
private const CHUNK_SIZE = 1000; // Process data in chunks to reduce memory usage
|
||||||
|
|
||||||
private string $period = '';
|
private string $period = '';
|
||||||
private int $processedCount = 0;
|
private int $processedCount = 0;
|
||||||
private int $errorCount = 0;
|
private int $errorCount = 0;
|
||||||
|
private array $billDetailBatch = [];
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Create a new job instance.
|
* Create a new job instance.
|
||||||
@@ -61,6 +63,7 @@
|
|||||||
set_time_limit(self::MAX_EXECUTION_TIME);
|
set_time_limit(self::MAX_EXECUTION_TIME);
|
||||||
$this->processedCount = 0;
|
$this->processedCount = 0;
|
||||||
$this->errorCount = 0;
|
$this->errorCount = 0;
|
||||||
|
$this->billDetailBatch = [];
|
||||||
}
|
}
|
||||||
|
|
||||||
private function processPeriod()
|
private function processPeriod()
|
||||||
@@ -111,10 +114,23 @@
|
|||||||
|
|
||||||
$headers = (new TempBillDetail())->getFillable();
|
$headers = (new TempBillDetail())->getFillable();
|
||||||
$rowCount = 0;
|
$rowCount = 0;
|
||||||
|
$chunkCount = 0;
|
||||||
|
|
||||||
while (($row = fgetcsv($handle, 0, self::CSV_DELIMITER)) !== false) {
|
while (($row = fgetcsv($handle, 0, self::CSV_DELIMITER)) !== false) {
|
||||||
$rowCount++;
|
$rowCount++;
|
||||||
$this->processRow($row, $headers, $rowCount, $filePath);
|
$this->processRow($row, $headers, $rowCount, $filePath);
|
||||||
|
|
||||||
|
// Process in chunks to avoid memory issues
|
||||||
|
if (count($this->billDetailBatch) >= self::CHUNK_SIZE) {
|
||||||
|
$this->saveBatch();
|
||||||
|
$chunkCount++;
|
||||||
|
Log::info("Processed chunk $chunkCount ({$this->processedCount} records so far)");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Process any remaining records
|
||||||
|
if (!empty($this->billDetailBatch)) {
|
||||||
|
$this->saveBatch();
|
||||||
}
|
}
|
||||||
|
|
||||||
fclose($handle);
|
fclose($handle);
|
||||||
@@ -127,22 +143,29 @@
|
|||||||
if (count($headers) !== count($row)) {
|
if (count($headers) !== count($row)) {
|
||||||
Log::warning("Row $rowCount in $filePath has incorrect column count. Expected: " .
|
Log::warning("Row $rowCount in $filePath has incorrect column count. Expected: " .
|
||||||
count($headers) . ", Got: " . count($row));
|
count($headers) . ", Got: " . count($row));
|
||||||
|
$this->errorCount++;
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
$data = array_combine($headers, $row);
|
$data = array_combine($headers, $row);
|
||||||
$this->saveRecord($data, $rowCount, $filePath);
|
$this->addToBatch($data, $rowCount, $filePath);
|
||||||
}
|
}
|
||||||
|
|
||||||
private function saveRecord(array $data, int $rowCount, string $filePath)
|
/**
|
||||||
|
* Add record to batch instead of saving immediately
|
||||||
|
*/
|
||||||
|
private function addToBatch(array $data, int $rowCount, string $filePath)
|
||||||
: void
|
: void
|
||||||
{
|
{
|
||||||
try {
|
try {
|
||||||
if (isset($data['_id']) && $data['_id'] !== '_id') {
|
if (isset($data['_id']) && $data['_id'] !== '_id') {
|
||||||
TempBillDetail::updateOrCreate(
|
// Add timestamp fields
|
||||||
['_id' => $data['_id']],
|
$now = now();
|
||||||
$data
|
$data['created_at'] = $now;
|
||||||
);
|
$data['updated_at'] = $now;
|
||||||
|
|
||||||
|
// Add to batch
|
||||||
|
$this->billDetailBatch[] = $data;
|
||||||
$this->processedCount++;
|
$this->processedCount++;
|
||||||
}
|
}
|
||||||
} catch (Exception $e) {
|
} catch (Exception $e) {
|
||||||
@@ -151,6 +174,32 @@
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Save batched records to the database
|
||||||
|
*/
|
||||||
|
private function saveBatch()
|
||||||
|
: void
|
||||||
|
{
|
||||||
|
try {
|
||||||
|
if (!empty($this->billDetailBatch)) {
|
||||||
|
// Bulk insert/update bill details
|
||||||
|
TempBillDetail::upsert(
|
||||||
|
$this->billDetailBatch,
|
||||||
|
['_id'], // Unique key
|
||||||
|
array_diff((new TempBillDetail())->getFillable(), ['_id']) // Update columns
|
||||||
|
);
|
||||||
|
|
||||||
|
// Reset batch after processing
|
||||||
|
$this->billDetailBatch = [];
|
||||||
|
}
|
||||||
|
} catch (Exception $e) {
|
||||||
|
Log::error("Error in saveBatch: " . $e->getMessage());
|
||||||
|
$this->errorCount += count($this->billDetailBatch);
|
||||||
|
// Reset batch even if there's an error to prevent reprocessing the same failed records
|
||||||
|
$this->billDetailBatch = [];
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
private function cleanup(string $tempFilePath)
|
private function cleanup(string $tempFilePath)
|
||||||
: void
|
: void
|
||||||
{
|
{
|
||||||
|
|||||||
@@ -20,10 +20,12 @@
|
|||||||
private const MAX_EXECUTION_TIME = 86400; // 24 hours in seconds
|
private const MAX_EXECUTION_TIME = 86400; // 24 hours in seconds
|
||||||
private const FILENAME = 'ST.CUSTOMER.csv';
|
private const FILENAME = 'ST.CUSTOMER.csv';
|
||||||
private const DISK_NAME = 'sftpStatement';
|
private const DISK_NAME = 'sftpStatement';
|
||||||
|
private const CHUNK_SIZE = 1000; // Process data in chunks to reduce memory usage
|
||||||
|
|
||||||
private string $period = '';
|
private string $period = '';
|
||||||
private int $processedCount = 0;
|
private int $processedCount = 0;
|
||||||
private int $errorCount = 0;
|
private int $errorCount = 0;
|
||||||
|
private array $customerBatch = [];
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Create a new job instance.
|
* Create a new job instance.
|
||||||
@@ -61,6 +63,7 @@
|
|||||||
set_time_limit(self::MAX_EXECUTION_TIME);
|
set_time_limit(self::MAX_EXECUTION_TIME);
|
||||||
$this->processedCount = 0;
|
$this->processedCount = 0;
|
||||||
$this->errorCount = 0;
|
$this->errorCount = 0;
|
||||||
|
$this->customerBatch = [];
|
||||||
}
|
}
|
||||||
|
|
||||||
private function processPeriod()
|
private function processPeriod()
|
||||||
@@ -111,10 +114,23 @@
|
|||||||
|
|
||||||
$headers = (new Customer())->getFillable();
|
$headers = (new Customer())->getFillable();
|
||||||
$rowCount = 0;
|
$rowCount = 0;
|
||||||
|
$chunkCount = 0;
|
||||||
|
|
||||||
while (($row = fgetcsv($handle, 0, self::CSV_DELIMITER)) !== false) {
|
while (($row = fgetcsv($handle, 0, self::CSV_DELIMITER)) !== false) {
|
||||||
$rowCount++;
|
$rowCount++;
|
||||||
$this->processRow($row, $headers, $rowCount, $filePath);
|
$this->processRow($row, $headers, $rowCount, $filePath);
|
||||||
|
|
||||||
|
// Process in chunks to avoid memory issues
|
||||||
|
if (count($this->customerBatch) >= self::CHUNK_SIZE) {
|
||||||
|
$this->saveBatch();
|
||||||
|
$chunkCount++;
|
||||||
|
Log::info("Processed chunk $chunkCount ({$this->processedCount} records so far)");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Process any remaining records
|
||||||
|
if (!empty($this->customerBatch)) {
|
||||||
|
$this->saveBatch();
|
||||||
}
|
}
|
||||||
|
|
||||||
fclose($handle);
|
fclose($handle);
|
||||||
@@ -131,17 +147,24 @@
|
|||||||
}
|
}
|
||||||
|
|
||||||
$data = array_combine($headers, $row);
|
$data = array_combine($headers, $row);
|
||||||
$this->saveRecord($data, $rowCount, $filePath);
|
$this->addToBatch($data, $rowCount, $filePath);
|
||||||
}
|
}
|
||||||
|
|
||||||
private function saveRecord(array $data, int $rowCount, string $filePath)
|
/**
|
||||||
|
* Add record to batch instead of saving immediately
|
||||||
|
*/
|
||||||
|
private function addToBatch(array $data, int $rowCount, string $filePath)
|
||||||
: void
|
: void
|
||||||
{
|
{
|
||||||
try {
|
try {
|
||||||
if (isset($data['customer_code']) && $data['customer_code'] !== 'customer_code') {
|
if (isset($data['customer_code']) && $data['customer_code'] !== 'customer_code') {
|
||||||
$customer = Customer::firstOrNew(['customer_code' => $data['customer_code']]);
|
// Add timestamp fields
|
||||||
$customer->fill($data);
|
$now = now();
|
||||||
$customer->save();
|
$data['created_at'] = $now;
|
||||||
|
$data['updated_at'] = $now;
|
||||||
|
|
||||||
|
// Add to customer batch
|
||||||
|
$this->customerBatch[] = $data;
|
||||||
$this->processedCount++;
|
$this->processedCount++;
|
||||||
}
|
}
|
||||||
} catch (Exception $e) {
|
} catch (Exception $e) {
|
||||||
@@ -150,6 +173,32 @@
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Save batched records to the database
|
||||||
|
*/
|
||||||
|
private function saveBatch()
|
||||||
|
: void
|
||||||
|
{
|
||||||
|
try {
|
||||||
|
if (!empty($this->customerBatch)) {
|
||||||
|
// Bulk insert/update customers
|
||||||
|
Customer::upsert(
|
||||||
|
$this->customerBatch,
|
||||||
|
['customer_code'], // Unique key
|
||||||
|
array_diff((new Customer())->getFillable(), ['customer_code']) // Update columns
|
||||||
|
);
|
||||||
|
|
||||||
|
// Reset customer batch after processing
|
||||||
|
$this->customerBatch = [];
|
||||||
|
}
|
||||||
|
} catch (Exception $e) {
|
||||||
|
Log::error("Error in saveBatch: " . $e->getMessage());
|
||||||
|
$this->errorCount += count($this->customerBatch);
|
||||||
|
// Reset batch even if there's an error to prevent reprocessing the same failed records
|
||||||
|
$this->customerBatch = [];
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
private function cleanup(string $tempFilePath)
|
private function cleanup(string $tempFilePath)
|
||||||
: void
|
: void
|
||||||
{
|
{
|
||||||
|
|||||||
@@ -20,6 +20,7 @@
|
|||||||
private const MAX_EXECUTION_TIME = 86400; // 24 hours in seconds
|
private const MAX_EXECUTION_TIME = 86400; // 24 hours in seconds
|
||||||
private const FILENAME = 'ST.DATA.CAPTURE.csv';
|
private const FILENAME = 'ST.DATA.CAPTURE.csv';
|
||||||
private const DISK_NAME = 'sftpStatement';
|
private const DISK_NAME = 'sftpStatement';
|
||||||
|
private const CHUNK_SIZE = 1000; // Process data in chunks to reduce memory usage
|
||||||
private const CSV_HEADERS = [
|
private const CSV_HEADERS = [
|
||||||
'id',
|
'id',
|
||||||
'account_number',
|
'account_number',
|
||||||
@@ -66,6 +67,7 @@
|
|||||||
private string $period = '';
|
private string $period = '';
|
||||||
private int $processedCount = 0;
|
private int $processedCount = 0;
|
||||||
private int $errorCount = 0;
|
private int $errorCount = 0;
|
||||||
|
private array $captureBatch = [];
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Create a new job instance.
|
* Create a new job instance.
|
||||||
@@ -103,6 +105,7 @@
|
|||||||
set_time_limit(self::MAX_EXECUTION_TIME);
|
set_time_limit(self::MAX_EXECUTION_TIME);
|
||||||
$this->processedCount = 0;
|
$this->processedCount = 0;
|
||||||
$this->errorCount = 0;
|
$this->errorCount = 0;
|
||||||
|
$this->captureBatch = [];
|
||||||
}
|
}
|
||||||
|
|
||||||
private function processPeriod()
|
private function processPeriod()
|
||||||
@@ -152,6 +155,8 @@
|
|||||||
}
|
}
|
||||||
|
|
||||||
$rowCount = 0;
|
$rowCount = 0;
|
||||||
|
$chunkCount = 0;
|
||||||
|
|
||||||
while (($row = fgetcsv($handle, 0, self::CSV_DELIMITER)) !== false) {
|
while (($row = fgetcsv($handle, 0, self::CSV_DELIMITER)) !== false) {
|
||||||
$rowCount++;
|
$rowCount++;
|
||||||
|
|
||||||
@@ -161,6 +166,18 @@
|
|||||||
}
|
}
|
||||||
|
|
||||||
$this->processRow($row, $rowCount, $filePath);
|
$this->processRow($row, $rowCount, $filePath);
|
||||||
|
|
||||||
|
// Process in chunks to avoid memory issues
|
||||||
|
if (count($this->captureBatch) >= self::CHUNK_SIZE) {
|
||||||
|
$this->saveBatch();
|
||||||
|
$chunkCount++;
|
||||||
|
Log::info("Processed chunk $chunkCount ({$this->processedCount} records so far)");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Process any remaining records
|
||||||
|
if (!empty($this->captureBatch)) {
|
||||||
|
$this->saveBatch();
|
||||||
}
|
}
|
||||||
|
|
||||||
fclose($handle);
|
fclose($handle);
|
||||||
@@ -178,7 +195,7 @@
|
|||||||
|
|
||||||
$data = array_combine(self::CSV_HEADERS, $row);
|
$data = array_combine(self::CSV_HEADERS, $row);
|
||||||
$this->formatDates($data);
|
$this->formatDates($data);
|
||||||
$this->saveRecord($data, $rowCount, $filePath);
|
$this->addToBatch($data, $rowCount, $filePath);
|
||||||
}
|
}
|
||||||
|
|
||||||
private function formatDates(array &$data)
|
private function formatDates(array &$data)
|
||||||
@@ -207,15 +224,21 @@
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
private function saveRecord(array $data, int $rowCount, string $filePath)
|
/**
|
||||||
|
* Add record to batch instead of saving immediately
|
||||||
|
*/
|
||||||
|
private function addToBatch(array $data, int $rowCount, string $filePath)
|
||||||
: void
|
: void
|
||||||
{
|
{
|
||||||
try {
|
try {
|
||||||
if (!empty($data['id'])) {
|
if (!empty($data['id'])) {
|
||||||
DataCapture::updateOrCreate(
|
// Add timestamp fields
|
||||||
['id' => $data['id']],
|
$now = now();
|
||||||
$data
|
$data['created_at'] = $now;
|
||||||
);
|
$data['updated_at'] = $now;
|
||||||
|
|
||||||
|
// Add to capture batch
|
||||||
|
$this->captureBatch[] = $data;
|
||||||
$this->processedCount++;
|
$this->processedCount++;
|
||||||
}
|
}
|
||||||
} catch (Exception $e) {
|
} catch (Exception $e) {
|
||||||
@@ -224,6 +247,32 @@
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Save batched records to the database
|
||||||
|
*/
|
||||||
|
private function saveBatch()
|
||||||
|
: void
|
||||||
|
{
|
||||||
|
try {
|
||||||
|
if (!empty($this->captureBatch)) {
|
||||||
|
// Bulk insert/update data captures
|
||||||
|
DataCapture::upsert(
|
||||||
|
$this->captureBatch,
|
||||||
|
['id'], // Unique key
|
||||||
|
array_diff(self::CSV_HEADERS, ['id']) // Update columns
|
||||||
|
);
|
||||||
|
|
||||||
|
// Reset capture batch after processing
|
||||||
|
$this->captureBatch = [];
|
||||||
|
}
|
||||||
|
} catch (Exception $e) {
|
||||||
|
Log::error("Error in saveBatch: " . $e->getMessage());
|
||||||
|
$this->errorCount += count($this->captureBatch);
|
||||||
|
// Reset batch even if there's an error to prevent reprocessing the same failed records
|
||||||
|
$this->captureBatch = [];
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
private function cleanup(string $tempFilePath)
|
private function cleanup(string $tempFilePath)
|
||||||
: void
|
: void
|
||||||
{
|
{
|
||||||
|
|||||||
@@ -20,10 +20,12 @@
|
|||||||
private const MAX_EXECUTION_TIME = 86400; // 24 hours in seconds
|
private const MAX_EXECUTION_TIME = 86400; // 24 hours in seconds
|
||||||
private const FILENAME = 'ST.FUNDS.TRANSFER.csv';
|
private const FILENAME = 'ST.FUNDS.TRANSFER.csv';
|
||||||
private const DISK_NAME = 'sftpStatement';
|
private const DISK_NAME = 'sftpStatement';
|
||||||
|
private const CHUNK_SIZE = 1000; // Process data in chunks to reduce memory usage
|
||||||
|
|
||||||
private string $period = '';
|
private string $period = '';
|
||||||
private int $processedCount = 0;
|
private int $processedCount = 0;
|
||||||
private int $errorCount = 0;
|
private int $errorCount = 0;
|
||||||
|
private array $transferBatch = [];
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Create a new job instance.
|
* Create a new job instance.
|
||||||
@@ -61,6 +63,7 @@
|
|||||||
set_time_limit(self::MAX_EXECUTION_TIME);
|
set_time_limit(self::MAX_EXECUTION_TIME);
|
||||||
$this->processedCount = 0;
|
$this->processedCount = 0;
|
||||||
$this->errorCount = 0;
|
$this->errorCount = 0;
|
||||||
|
$this->transferBatch = [];
|
||||||
}
|
}
|
||||||
|
|
||||||
private function processPeriod()
|
private function processPeriod()
|
||||||
@@ -111,10 +114,23 @@
|
|||||||
|
|
||||||
$headers = (new TempFundsTransfer())->getFillable();
|
$headers = (new TempFundsTransfer())->getFillable();
|
||||||
$rowCount = 0;
|
$rowCount = 0;
|
||||||
|
$chunkCount = 0;
|
||||||
|
|
||||||
while (($row = fgetcsv($handle, 0, self::CSV_DELIMITER)) !== false) {
|
while (($row = fgetcsv($handle, 0, self::CSV_DELIMITER)) !== false) {
|
||||||
$rowCount++;
|
$rowCount++;
|
||||||
$this->processRow($row, $headers, $rowCount, $filePath);
|
$this->processRow($row, $headers, $rowCount, $filePath);
|
||||||
|
|
||||||
|
// Process in chunks to avoid memory issues
|
||||||
|
if (count($this->transferBatch) >= self::CHUNK_SIZE) {
|
||||||
|
$this->saveBatch();
|
||||||
|
$chunkCount++;
|
||||||
|
Log::info("Processed chunk $chunkCount ({$this->processedCount} records so far)");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Process any remaining records
|
||||||
|
if (!empty($this->transferBatch)) {
|
||||||
|
$this->saveBatch();
|
||||||
}
|
}
|
||||||
|
|
||||||
fclose($handle);
|
fclose($handle);
|
||||||
@@ -136,18 +152,24 @@
|
|||||||
}
|
}
|
||||||
|
|
||||||
$data = array_combine($headers, $row);
|
$data = array_combine($headers, $row);
|
||||||
$this->saveRecord($data, $rowCount, $filePath);
|
$this->addToBatch($data, $rowCount, $filePath);
|
||||||
}
|
}
|
||||||
|
|
||||||
private function saveRecord(array $data, int $rowCount, string $filePath)
|
/**
|
||||||
|
* Add record to batch instead of saving immediately
|
||||||
|
*/
|
||||||
|
private function addToBatch(array $data, int $rowCount, string $filePath)
|
||||||
: void
|
: void
|
||||||
{
|
{
|
||||||
try {
|
try {
|
||||||
if (isset($data['_id']) && $data['_id'] !== '_id') {
|
if (isset($data['_id']) && $data['_id'] !== '_id') {
|
||||||
TempFundsTransfer::updateOrCreate(
|
// Add timestamp fields
|
||||||
['_id' => $data['_id']],
|
$now = now();
|
||||||
$data
|
$data['created_at'] = $now;
|
||||||
);
|
$data['updated_at'] = $now;
|
||||||
|
|
||||||
|
// Add to transfer batch
|
||||||
|
$this->transferBatch[] = $data;
|
||||||
$this->processedCount++;
|
$this->processedCount++;
|
||||||
}
|
}
|
||||||
} catch (Exception $e) {
|
} catch (Exception $e) {
|
||||||
@@ -156,6 +178,32 @@
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Save batched records to the database
|
||||||
|
*/
|
||||||
|
private function saveBatch()
|
||||||
|
: void
|
||||||
|
{
|
||||||
|
try {
|
||||||
|
if (!empty($this->transferBatch)) {
|
||||||
|
// Bulk insert/update funds transfers
|
||||||
|
TempFundsTransfer::upsert(
|
||||||
|
$this->transferBatch,
|
||||||
|
['_id'], // Unique key
|
||||||
|
array_diff((new TempFundsTransfer())->getFillable(), ['_id']) // Update columns
|
||||||
|
);
|
||||||
|
|
||||||
|
// Reset transfer batch after processing
|
||||||
|
$this->transferBatch = [];
|
||||||
|
}
|
||||||
|
} catch (Exception $e) {
|
||||||
|
Log::error("Error in saveBatch: " . $e->getMessage());
|
||||||
|
$this->errorCount += count($this->transferBatch);
|
||||||
|
// Reset batch even if there's an error to prevent reprocessing the same failed records
|
||||||
|
$this->transferBatch = [];
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
private function cleanup(string $tempFilePath)
|
private function cleanup(string $tempFilePath)
|
||||||
: void
|
: void
|
||||||
{
|
{
|
||||||
|
|||||||
@@ -20,10 +20,12 @@
|
|||||||
private const MAX_EXECUTION_TIME = 86400; // 24 hours in seconds
|
private const MAX_EXECUTION_TIME = 86400; // 24 hours in seconds
|
||||||
private const FILENAME = 'ST.STMT.ENTRY.csv';
|
private const FILENAME = 'ST.STMT.ENTRY.csv';
|
||||||
private const DISK_NAME = 'sftpStatement';
|
private const DISK_NAME = 'sftpStatement';
|
||||||
|
private const CHUNK_SIZE = 1000; // Process data in chunks to reduce memory usage
|
||||||
|
|
||||||
private string $period = '';
|
private string $period = '';
|
||||||
private int $processedCount = 0;
|
private int $processedCount = 0;
|
||||||
private int $errorCount = 0;
|
private int $errorCount = 0;
|
||||||
|
private array $entryBatch = [];
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Create a new job instance.
|
* Create a new job instance.
|
||||||
@@ -61,6 +63,7 @@
|
|||||||
set_time_limit(self::MAX_EXECUTION_TIME);
|
set_time_limit(self::MAX_EXECUTION_TIME);
|
||||||
$this->processedCount = 0;
|
$this->processedCount = 0;
|
||||||
$this->errorCount = 0;
|
$this->errorCount = 0;
|
||||||
|
$this->entryBatch = [];
|
||||||
}
|
}
|
||||||
|
|
||||||
private function processPeriod()
|
private function processPeriod()
|
||||||
@@ -111,10 +114,23 @@
|
|||||||
|
|
||||||
$headers = (new StmtEntry())->getFillable();
|
$headers = (new StmtEntry())->getFillable();
|
||||||
$rowCount = 0;
|
$rowCount = 0;
|
||||||
|
$chunkCount = 0;
|
||||||
|
|
||||||
while (($row = fgetcsv($handle, 0, self::CSV_DELIMITER)) !== false) {
|
while (($row = fgetcsv($handle, 0, self::CSV_DELIMITER)) !== false) {
|
||||||
$rowCount++;
|
$rowCount++;
|
||||||
$this->processRow($row, $headers, $rowCount, $filePath);
|
$this->processRow($row, $headers, $rowCount, $filePath);
|
||||||
|
|
||||||
|
// Process in chunks to avoid memory issues
|
||||||
|
if (count($this->entryBatch) >= self::CHUNK_SIZE) {
|
||||||
|
$this->saveBatch();
|
||||||
|
$chunkCount++;
|
||||||
|
Log::info("Processed chunk $chunkCount ({$this->processedCount} records so far)");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Process any remaining records
|
||||||
|
if (!empty($this->entryBatch)) {
|
||||||
|
$this->saveBatch();
|
||||||
}
|
}
|
||||||
|
|
||||||
fclose($handle);
|
fclose($handle);
|
||||||
@@ -132,7 +148,7 @@
|
|||||||
|
|
||||||
$data = array_combine($headers, $row);
|
$data = array_combine($headers, $row);
|
||||||
$this->cleanTransReference($data);
|
$this->cleanTransReference($data);
|
||||||
$this->saveRecord($data, $rowCount, $filePath);
|
$this->addToBatch($data, $rowCount, $filePath);
|
||||||
}
|
}
|
||||||
|
|
||||||
private function cleanTransReference(array &$data)
|
private function cleanTransReference(array &$data)
|
||||||
@@ -144,15 +160,21 @@
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
private function saveRecord(array $data, int $rowCount, string $filePath)
|
/**
|
||||||
|
* Add record to batch instead of saving immediately
|
||||||
|
*/
|
||||||
|
private function addToBatch(array $data, int $rowCount, string $filePath)
|
||||||
: void
|
: void
|
||||||
{
|
{
|
||||||
try {
|
try {
|
||||||
if (isset($data['stmt_entry_id']) && $data['stmt_entry_id'] !== 'stmt_entry_id') {
|
if (isset($data['stmt_entry_id']) && $data['stmt_entry_id'] !== 'stmt_entry_id') {
|
||||||
StmtEntry::updateOrCreate(
|
// Add timestamp fields
|
||||||
['stmt_entry_id' => $data['stmt_entry_id']],
|
$now = now();
|
||||||
$data
|
$data['created_at'] = $now;
|
||||||
);
|
$data['updated_at'] = $now;
|
||||||
|
|
||||||
|
// Add to entry batch
|
||||||
|
$this->entryBatch[] = $data;
|
||||||
$this->processedCount++;
|
$this->processedCount++;
|
||||||
}
|
}
|
||||||
} catch (Exception $e) {
|
} catch (Exception $e) {
|
||||||
@@ -161,6 +183,32 @@
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Save batched records to the database
|
||||||
|
*/
|
||||||
|
private function saveBatch()
|
||||||
|
: void
|
||||||
|
{
|
||||||
|
try {
|
||||||
|
if (!empty($this->entryBatch)) {
|
||||||
|
// Bulk insert/update statement entries
|
||||||
|
StmtEntry::upsert(
|
||||||
|
$this->entryBatch,
|
||||||
|
['stmt_entry_id'], // Unique key
|
||||||
|
array_diff((new StmtEntry())->getFillable(), ['stmt_entry_id']) // Update columns
|
||||||
|
);
|
||||||
|
|
||||||
|
// Reset entry batch after processing
|
||||||
|
$this->entryBatch = [];
|
||||||
|
}
|
||||||
|
} catch (Exception $e) {
|
||||||
|
Log::error("Error in saveBatch: " . $e->getMessage());
|
||||||
|
$this->errorCount += count($this->entryBatch);
|
||||||
|
// Reset batch even if there's an error to prevent reprocessing the same failed records
|
||||||
|
$this->entryBatch = [];
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
private function cleanup(string $tempFilePath)
|
private function cleanup(string $tempFilePath)
|
||||||
: void
|
: void
|
||||||
{
|
{
|
||||||
|
|||||||
@@ -20,6 +20,7 @@
|
|||||||
private const MAX_EXECUTION_TIME = 86400; // 24 hours in seconds
|
private const MAX_EXECUTION_TIME = 86400; // 24 hours in seconds
|
||||||
private const FILENAME = 'ST.TELLER.csv';
|
private const FILENAME = 'ST.TELLER.csv';
|
||||||
private const DISK_NAME = 'sftpStatement';
|
private const DISK_NAME = 'sftpStatement';
|
||||||
|
private const CHUNK_SIZE = 1000; // Process data in chunks to reduce memory usage
|
||||||
private const HEADER_MAP = [
|
private const HEADER_MAP = [
|
||||||
'id' => 'id_teller',
|
'id' => 'id_teller',
|
||||||
'account_1' => 'account_1',
|
'account_1' => 'account_1',
|
||||||
@@ -129,6 +130,7 @@
|
|||||||
private string $period = '';
|
private string $period = '';
|
||||||
private int $processedCount = 0;
|
private int $processedCount = 0;
|
||||||
private int $errorCount = 0;
|
private int $errorCount = 0;
|
||||||
|
private array $tellerBatch = [];
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Create a new job instance.
|
* Create a new job instance.
|
||||||
@@ -166,6 +168,7 @@
|
|||||||
set_time_limit(self::MAX_EXECUTION_TIME);
|
set_time_limit(self::MAX_EXECUTION_TIME);
|
||||||
$this->processedCount = 0;
|
$this->processedCount = 0;
|
||||||
$this->errorCount = 0;
|
$this->errorCount = 0;
|
||||||
|
$this->tellerBatch = [];
|
||||||
}
|
}
|
||||||
|
|
||||||
private function processPeriod()
|
private function processPeriod()
|
||||||
@@ -222,9 +225,23 @@
|
|||||||
}
|
}
|
||||||
|
|
||||||
$rowCount = 0;
|
$rowCount = 0;
|
||||||
|
$chunkCount = 0;
|
||||||
|
|
||||||
while (($row = fgetcsv($handle, 0, self::CSV_DELIMITER)) !== false) {
|
while (($row = fgetcsv($handle, 0, self::CSV_DELIMITER)) !== false) {
|
||||||
$rowCount++;
|
$rowCount++;
|
||||||
$this->processRow($headerRow, $row, $rowCount, $filePath);
|
$this->processRow($headerRow, $row, $rowCount, $filePath);
|
||||||
|
|
||||||
|
// Process in chunks to avoid memory issues
|
||||||
|
if (count($this->tellerBatch) >= self::CHUNK_SIZE) {
|
||||||
|
$this->saveBatch();
|
||||||
|
$chunkCount++;
|
||||||
|
Log::info("Processed chunk $chunkCount ({$this->processedCount} records so far)");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Process any remaining records
|
||||||
|
if (!empty($this->tellerBatch)) {
|
||||||
|
$this->saveBatch();
|
||||||
}
|
}
|
||||||
|
|
||||||
fclose($handle);
|
fclose($handle);
|
||||||
@@ -234,6 +251,14 @@
|
|||||||
private function processRow(array $headerRow, array $row, int $rowCount, string $filePath)
|
private function processRow(array $headerRow, array $row, int $rowCount, string $filePath)
|
||||||
: void
|
: void
|
||||||
{
|
{
|
||||||
|
// Skip if row doesn't have enough columns
|
||||||
|
if (count($headerRow) !== count($row)) {
|
||||||
|
Log::warning("Row $rowCount in $filePath has incorrect column count. Expected: " .
|
||||||
|
count($headerRow) . ", Got: " . count($row));
|
||||||
|
$this->errorCount++;
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
// Combine the header row with the data row
|
// Combine the header row with the data row
|
||||||
$rawData = array_combine($headerRow, $row);
|
$rawData = array_combine($headerRow, $row);
|
||||||
|
|
||||||
@@ -249,10 +274,13 @@
|
|||||||
}
|
}
|
||||||
|
|
||||||
try {
|
try {
|
||||||
$teller = Teller::firstOrNew(['id_teller' => $data['id_teller']]);
|
// Add timestamps
|
||||||
$teller->fill($data);
|
$now = now();
|
||||||
$teller->save();
|
$data['created_at'] = $now;
|
||||||
|
$data['updated_at'] = $now;
|
||||||
|
|
||||||
|
// Add to batch for bulk processing
|
||||||
|
$this->tellerBatch[] = $data;
|
||||||
$this->processedCount++;
|
$this->processedCount++;
|
||||||
} catch (Exception $e) {
|
} catch (Exception $e) {
|
||||||
Log::error("Error processing Teller at row $rowCount in $filePath: " . $e->getMessage());
|
Log::error("Error processing Teller at row $rowCount in $filePath: " . $e->getMessage());
|
||||||
@@ -260,6 +288,28 @@
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
private function saveBatch(): void
|
||||||
|
{
|
||||||
|
try {
|
||||||
|
if (!empty($this->tellerBatch)) {
|
||||||
|
// Bulk insert/update teller records
|
||||||
|
Teller::upsert(
|
||||||
|
$this->tellerBatch,
|
||||||
|
['id_teller'], // Unique key
|
||||||
|
array_diff(array_values(self::HEADER_MAP), ['id_teller']) // Update columns
|
||||||
|
);
|
||||||
|
|
||||||
|
// Reset batch after processing
|
||||||
|
$this->tellerBatch = [];
|
||||||
|
}
|
||||||
|
} catch (Exception $e) {
|
||||||
|
Log::error("Error in saveBatch: " . $e->getMessage());
|
||||||
|
$this->errorCount += count($this->tellerBatch);
|
||||||
|
// Reset batch even if there's an error to prevent reprocessing the same failed records
|
||||||
|
$this->tellerBatch = [];
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
private function logJobCompletion()
|
private function logJobCompletion()
|
||||||
: void
|
: void
|
||||||
{
|
{
|
||||||
|
|||||||
Reference in New Issue
Block a user