MediaWiki  REL1_21
FileOpBatch.php
Go to the documentation of this file.
00001 <?php
00034 class FileOpBatch {
00035         /* Timeout related parameters */
00036         const MAX_BATCH_SIZE = 1000; // integer
00037 
00057         public static function attempt( array $performOps, array $opts, FileJournal $journal ) {
00058                 wfProfileIn( __METHOD__ );
00059                 $status = Status::newGood();
00060 
00061                 $n = count( $performOps );
00062                 if ( $n > self::MAX_BATCH_SIZE ) {
00063                         $status->fatal( 'backend-fail-batchsize', $n, self::MAX_BATCH_SIZE );
00064                         wfProfileOut( __METHOD__ );
00065                         return $status;
00066                 }
00067 
00068                 $batchId = $journal->getTimestampedUUID();
00069                 $ignoreErrors = !empty( $opts['force'] );
00070                 $journaled = empty( $opts['nonJournaled'] );
00071                 $maxConcurrency = isset( $opts['concurrency'] ) ? $opts['concurrency'] : 1;
00072 
00073                 $entries = array(); // file journal entry list
00074                 $predicates = FileOp::newPredicates(); // account for previous ops in prechecks
00075                 $curBatch = array(); // concurrent FileOp sub-batch accumulation
00076                 $curBatchDeps = FileOp::newDependencies(); // paths used in FileOp sub-batch
00077                 $pPerformOps = array(); // ordered list of concurrent FileOp sub-batches
00078                 $lastBackend = null; // last op backend name
00079                 // Do pre-checks for each operation; abort on failure...
00080                 foreach ( $performOps as $index => $fileOp ) {
00081                         $backendName = $fileOp->getBackend()->getName();
00082                         $fileOp->setBatchId( $batchId ); // transaction ID
00083                         // Decide if this op can be done concurrently within this sub-batch
00084                         // or if a new concurrent sub-batch must be started after this one...
00085                         if ( $fileOp->dependsOn( $curBatchDeps )
00086                                 || count( $curBatch ) >= $maxConcurrency
00087                                 || ( $backendName !== $lastBackend && count( $curBatch ) )
00088                         ) {
00089                                 $pPerformOps[] = $curBatch; // push this batch
00090                                 $curBatch = array(); // start a new sub-batch
00091                                 $curBatchDeps = FileOp::newDependencies();
00092                         }
00093                         $lastBackend = $backendName;
00094                         $curBatch[$index] = $fileOp; // keep index
00095                         // Update list of affected paths in this batch
00096                         $curBatchDeps = $fileOp->applyDependencies( $curBatchDeps );
00097                         // Simulate performing the operation...
00098                         $oldPredicates = $predicates;
00099                         $subStatus = $fileOp->precheck( $predicates ); // updates $predicates
00100                         $status->merge( $subStatus );
00101                         if ( $subStatus->isOK() ) {
00102                                 if ( $journaled ) { // journal log entries
00103                                         $entries = array_merge( $entries,
00104                                                 $fileOp->getJournalEntries( $oldPredicates, $predicates ) );
00105                                 }
00106                         } else { // operation failed?
00107                                 $status->success[$index] = false;
00108                                 ++$status->failCount;
00109                                 if ( !$ignoreErrors ) {
00110                                         wfProfileOut( __METHOD__ );
00111                                         return $status; // abort
00112                                 }
00113                         }
00114                 }
00115                 // Push the last sub-batch
00116                 if ( count( $curBatch ) ) {
00117                         $pPerformOps[] = $curBatch;
00118                 }
00119 
00120                 // Log the operations in the file journal...
00121                 if ( count( $entries ) ) {
00122                         $subStatus = $journal->logChangeBatch( $entries, $batchId );
00123                         if ( !$subStatus->isOK() ) {
00124                                 wfProfileOut( __METHOD__ );
00125                                 return $subStatus; // abort
00126                         }
00127                 }
00128 
00129                 if ( $ignoreErrors ) { // treat precheck() fatals as mere warnings
00130                         $status->setResult( true, $status->value );
00131                 }
00132 
00133                 // Attempt each operation (in parallel if allowed and possible)...
00134                 self::runParallelBatches( $pPerformOps, $status );
00135 
00136                 wfProfileOut( __METHOD__ );
00137                 return $status;
00138         }
00139 
00152         protected static function runParallelBatches( array $pPerformOps, Status $status ) {
00153                 $aborted = false; // set to true on unexpected errors
00154                 foreach ( $pPerformOps as $performOpsBatch ) {
00155                         if ( $aborted ) { // check batch op abort flag...
00156                                 // We can't continue (even with $ignoreErrors) as $predicates is wrong.
00157                                 // Log the remaining ops as failed for recovery...
00158                                 foreach ( $performOpsBatch as $i => $fileOp ) {
00159                                         $performOpsBatch[$i]->logFailure( 'attempt_aborted' );
00160                                 }
00161                                 continue;
00162                         }
00163                         $statuses = array();
00164                         $opHandles = array();
00165                         // Get the backend; all sub-batch ops belong to a single backend
00166                         $backend = reset( $performOpsBatch )->getBackend();
00167                         // Get the operation handles or actually do it if there is just one.
00168                         // If attemptAsync() returns a Status, it was either due to an error
00169                         // or the backend does not support async ops and did it synchronously.
00170                         foreach ( $performOpsBatch as $i => $fileOp ) {
00171                                 if ( !$fileOp->failed() ) { // failed => already has Status
00172                                         // If the batch is just one operation, it's faster to avoid
00173                                         // pipelining as that can involve creating new TCP connections.
00174                                         $subStatus = ( count( $performOpsBatch ) > 1 )
00175                                                 ? $fileOp->attemptAsync()
00176                                                 : $fileOp->attempt();
00177                                         if ( $subStatus->value instanceof FileBackendStoreOpHandle ) {
00178                                                 $opHandles[$i] = $subStatus->value; // deferred
00179                                         } else {
00180                                                 $statuses[$i] = $subStatus; // done already
00181                                         }
00182                                 }
00183                         }
00184                         // Try to do all the operations concurrently...
00185                         $statuses = $statuses + $backend->executeOpHandlesInternal( $opHandles );
00186                         // Marshall and merge all the responses (blocking)...
00187                         foreach ( $performOpsBatch as $i => $fileOp ) {
00188                                 if ( !$fileOp->failed() ) { // failed => already has Status
00189                                         $subStatus = $statuses[$i];
00190                                         $status->merge( $subStatus );
00191                                         if ( $subStatus->isOK() ) {
00192                                                 $status->success[$i] = true;
00193                                                 ++$status->successCount;
00194                                         } else {
00195                                                 $status->success[$i] = false;
00196                                                 ++$status->failCount;
00197                                                 $aborted = true; // set abort flag; we can't continue
00198                                         }
00199                                 }
00200                         }
00201                 }
00202                 return $status;
00203         }
00204 }