X-Git-Url: http://git.cyclocoop.org/%7B%24www_url%7Dadmin/compta/pie.php?a=blobdiff_plain;f=includes%2Fapi%2FApiQueryDuplicateFiles.php;h=719c84a7b4fff94043b125119d7b33f1ba6ad13c;hb=6ef0901a4e75fb1118d6abe11fbed2374ffd29f7;hp=e05ffb6fae528b096714c730324c6bde5352d5fb;hpb=198a1a4dbda0fae34131e61fa7d18ab735c8c385;p=lhc%2Fweb%2Fwiklou.git diff --git a/includes/api/ApiQueryDuplicateFiles.php b/includes/api/ApiQueryDuplicateFiles.php index e05ffb6fae..719c84a7b4 100644 --- a/includes/api/ApiQueryDuplicateFiles.php +++ b/includes/api/ApiQueryDuplicateFiles.php @@ -59,73 +59,86 @@ class ApiQueryDuplicateFiles extends ApiQueryGeneratorBase { } $images = $namespaces[NS_FILE]; - $this->addTables( 'image', 'i1' ); - $this->addTables( 'image', 'i2' ); - $this->addFields( array( - 'i1.img_name AS orig_name', - 'i2.img_name AS dup_name', - 'i2.img_user_text AS dup_user_text', - 'i2.img_timestamp AS dup_timestamp' - ) ); - - $this->addWhere( array( - 'i1.img_name' => array_keys( $images ), - 'i1.img_sha1 = i2.img_sha1', - 'i1.img_name != i2.img_name', - ) ); + if( $params['dir'] == 'descending' ) { + $images = array_reverse( $images ); + } + $skipUntilThisDup = false; if ( isset( $params['continue'] ) ) { $cont = explode( '|', $params['continue'] ); if ( count( $cont ) != 2 ) { $this->dieUsage( 'Invalid continue param. You should pass the ' . 'original value returned by the previous query', '_badcontinue' ); } - $op = $params['dir'] == 'descending' ? '<' : '>'; - $db = $this->getDB(); - $orig = $db->addQuotes( $cont[0] ); - $dup = $db->addQuotes( $cont[1] ); - $this->addWhere( - "i1.img_name $op $orig OR " . - "(i1.img_name = $orig AND " . - "i2.img_name $op= $dup)" - ); + $fromImage = $cont[0]; + $skipUntilThisDup = $cont[1]; + // Filter out any images before $fromImage + foreach ( $images as $image => $pageId ) { + if ( $image < $fromImage ) { + unset( $images[$image] ); + } else { + break; + } + } } - $sort = ( $params['dir'] == 'descending' ? ' DESC' : '' ); - // Don't order by i1.img_name if it's constant in the WHERE clause - if ( count( $this->getPageSet()->getGoodTitles() ) == 1 ) { - $this->addOption( 'ORDER BY', 'i2.img_name' . $sort ); - } else { - $this->addOption( 'ORDER BY', array( - 'i1.img_name' . $sort, - 'i2.img_name' . $sort - )); - } - $this->addOption( 'LIMIT', $params['limit'] + 1 ); + $files = RepoGroup::singleton()->findFiles( array_keys( $images ) ); - $res = $this->select( __METHOD__ ); + $fit = true; $count = 0; $titles = array(); - foreach ( $res as $row ) { - if ( ++$count > $params['limit'] ) { - // We've reached the one extra which shows that - // there are additional pages to be had. Stop here... - $this->setContinueEnumParameter( 'continue', $row->orig_name . '|' . $row->dup_name ); - break; + + $sha1s = array(); + foreach ( $files as $file ) { + $sha1s[$file->getName()] = $file->getSha1(); + } + + // find all files with the hashes, result format is: array( hash => array( dup1, dup2 ), hash1 => ... ) + $filesBySha1s = RepoGroup::singleton()->findBySha1s( array_unique( array_values( $sha1s ) ) ); + + // iterate over $images to handle continue param correct + foreach( $images as $image => $pageId ) { + if( !isset( $sha1s[$image] ) ) { + continue; //file does not exist } - if ( !is_null( $resultPageSet ) ) { - $titles[] = Title::makeTitle( NS_FILE, $row->dup_name ); - } else { - $r = array( - 'name' => $row->dup_name, - 'user' => $row->dup_user_text, - 'timestamp' => wfTimestamp( TS_ISO_8601, $row->dup_timestamp ) - ); - $fit = $this->addPageSubItem( $images[$row->orig_name], $r ); - if ( !$fit ) { - $this->setContinueEnumParameter( 'continue', $row->orig_name . '|' . $row->dup_name ); + $sha1 = $sha1s[$image]; + $dupFiles = $filesBySha1s[$sha1]; + if( $params['dir'] == 'descending' ) { + $dupFiles = array_reverse( $dupFiles ); + } + foreach ( $dupFiles as $dupFile ) { + $dupName = $dupFile->getName(); + if( $image == $dupName ) { + continue; //ignore the file itself + } + if( $skipUntilThisDup !== false && $dupName < $skipUntilThisDup ) { + continue; //skip to pos after the image from continue param + } + $skipUntilThisDup = false; + if ( ++$count > $params['limit'] ) { + $fit = false; //break outer loop + // We're one over limit which shows that + // there are additional images to be had. Stop here... + $this->setContinueEnumParameter( 'continue', $image . '|' . $dupName ); break; } + if ( !is_null( $resultPageSet ) ) { + $titles[] = $file->getTitle(); + } else { + $r = array( + 'name' => $dupName, + 'user' => $dupFile->getUser( 'text' ), + 'timestamp' => wfTimestamp( TS_ISO_8601, $dupFile->getTimestamp() ) + ); + $fit = $this->addPageSubItem( $pageId, $r ); + if ( !$fit ) { + $this->setContinueEnumParameter( 'continue', $image . '|' . $dupName ); + break; + } + } + } + if( !$fit ) { + break; } } if ( !is_null( $resultPageSet ) ) { @@ -155,7 +168,7 @@ class ApiQueryDuplicateFiles extends ApiQueryGeneratorBase { public function getParamDescription() { return array( - 'limit' => 'How many files to return', + 'limit' => 'How many duplicate files to return', 'continue' => 'When more results are available, use this to continue', 'dir' => 'The direction in which to list', ); @@ -172,7 +185,7 @@ class ApiQueryDuplicateFiles extends ApiQueryGeneratorBase { } public function getDescription() { - return 'List all files that are duplicates of the given file(s)'; + return 'List all files that are duplicates of the given file(s) based on hash values'; } public function getPossibleErrors() {