disable(); $dbw = wfGetDB( DB_MASTER ); foreach ( $wgQueryPages as $page ) { @list( $class, $special, $limit ) = $page; # --list : just show the name of pages if( @$options['list'] ) { print "$special\n"; continue; } if ( $wgDisableQueryPageUpdate && in_array( $special, $wgDisableQueryPageUpdate ) ) { printf("%-30s disabled\n", $special); continue; } $specialObj = SpecialPage::getPage( $special ); if ( !$specialObj ) { print "No such special page: $special\n"; exit; } if ( !class_exists( $class ) ) { $file = $specialObj->getFile(); require_once( $file ); } $queryPage = new $class; if( !(isset($options['only'])) or ($options['only'] == $queryPage->getName()) ) { printf( '%-30s ', $special ); if ( $queryPage->isExpensive() ) { $t1 = explode( ' ', microtime() ); # Do the query $num = $queryPage->recache( $limit === null ? $wgQueryCacheLimit : $limit ); $t2 = explode( ' ', microtime() ); if ( $num === false ) { print "FAILED: database error\n"; } else { print "got $num rows in "; $elapsed = ($t2[0] - $t1[0]) + ($t2[1] - $t1[1]); $hours = intval( $elapsed / 3600 ); $minutes = intval( $elapsed % 3600 / 60 ); $seconds = $elapsed - $hours * 3600 - $minutes * 60; if ( $hours ) { print $hours . 'h '; } if ( $minutes ) { print $minutes . 'm '; } printf( "%.2fs\n", $seconds ); } # Reopen any connections that have closed if ( !$wgLoadBalancer->pingAll()) { print "\n"; do { print "Connection failed, reconnecting in 10 seconds...\n"; sleep(10); } while ( !$wgLoadBalancer->pingAll() ); print "Reconnected\n\n"; } else { # Commit the results $dbw->immediateCommit(); } # Wait for the slave to catch up /* $slaveDB = wfGetDB( DB_SLAVE, array('QueryPage::recache', 'vslow' ) ); while( $slaveDB->getLag() > 600 ) { print "Slave lagged, waiting...\n"; sleep(30); } */ wfWaitForSlaves( 5 ); } else { print "cheap, skipped\n"; } } }