X-Git-Url: http://git.cyclocoop.org/%22%20.%20generer_url_ecrire%28%22suivi_revisions%22%2C%22id_auteur=%24connecte%22%29%20.%20%22?a=blobdiff_plain;f=includes%2Fspecials%2FSpecialExport.php;h=3559ddee95acf9f0db2971d076850fb99ad129ce;hb=74a21f3bd1692dac958ddf3e09226a72b7bc65b7;hp=be58205f899f816bae21d7cb142d3ab5381bb7fd;hpb=0b6011413c8bf2c6d2f5bacff05b106dda3cb4f7;p=lhc%2Fweb%2Fwiklou.git diff --git a/includes/specials/SpecialExport.php b/includes/specials/SpecialExport.php index be58205f89..3559ddee95 100644 --- a/includes/specials/SpecialExport.php +++ b/includes/specials/SpecialExport.php @@ -22,40 +22,40 @@ */ class SpecialExport extends SpecialPage { - + private $curonly, $doExport, $pageLinkDepth, $templates; private $images; - + public function __construct() { parent::__construct( 'Export' ); } - + public function execute( $par ) { global $wgOut, $wgRequest, $wgSitename, $wgExportAllowListContributors; global $wgExportAllowHistory, $wgExportMaxHistory, $wgExportMaxLinkDepth; global $wgExportFromNamespaces; - + $this->setHeaders(); $this->outputHeader(); - + // Set some variables $this->curonly = true; $this->doExport = false; $this->templates = $wgRequest->getCheck( 'templates' ); $this->images = $wgRequest->getCheck( 'images' ); // Doesn't do anything yet $this->pageLinkDepth = $this->validateLinkDepth( - $wgRequest->getIntOrNull( 'pagelink-depth' ) ); + $wgRequest->getIntOrNull( 'pagelink-depth' ) ); $nsindex = ''; - + if ( $wgRequest->getCheck( 'addcat' ) ) { $page = $wgRequest->getText( 'pages' ); $catname = $wgRequest->getText( 'catname' ); - - if ( $catname !== '' && $catname !== NULL && $catname !== false ) { + + if ( $catname !== '' && $catname !== null && $catname !== false ) { $t = Title::makeTitleSafe( NS_MAIN, $catname ); if ( $t ) { /** - * @fixme This can lead to hitting memory limit for very large + * @todo Fixme: this can lead to hitting memory limit for very large * categories. Ideally we would do the lookup synchronously * during the export in a single query. */ @@ -67,14 +67,14 @@ class SpecialExport extends SpecialPage { else if( $wgRequest->getCheck( 'addns' ) && $wgExportFromNamespaces ) { $page = $wgRequest->getText( 'pages' ); $nsindex = $wgRequest->getText( 'nsindex', '' ); - - if ( $nsindex !== '' && $nsindex !== null && $nsindex !== false ) { + + if ( strval( $nsindex ) !== '' ) { /** - * Same implementation as above, so same @fixme + * Same implementation as above, so same @todo */ $nspages = $this->getPagesFromNamespace( $nsindex ); if ( $nspages ) $page .= "\n" . implode( "\n", $nspages ); - } + } } else if( $wgRequest->wasPosted() && $par == '' ) { $page = $wgRequest->getText( 'pages' ); @@ -88,15 +88,15 @@ class SpecialExport extends SpecialPage { $limit = $wgRequest->getInt( 'limit' ); $dir = $wgRequest->getVal( 'dir' ); $history = array( - 'dir' => 'asc', - 'offset' => false, - 'limit' => $wgExportMaxHistory, - ); + 'dir' => 'asc', + 'offset' => false, + 'limit' => $wgExportMaxHistory, + ); $historyCheck = $wgRequest->getCheck( 'history' ); if ( $this->curonly ) { $history = WikiExporter::CURRENT; } elseif ( !$historyCheck ) { - if ( $limit > 0 && $limit < $wgExportMaxHistory ) { + if ( $limit > 0 && ($wgExportMaxHistory == 0 || $limit < $wgExportMaxHistory ) ) { $history['limit'] = $limit; } if ( !is_null( $offset ) ) { @@ -106,7 +106,7 @@ class SpecialExport extends SpecialPage { $history['dir'] = 'desc'; } } - + if( $page != '' ) $this->doExport = true; } else { // Default to current-only for GET requests @@ -117,24 +117,24 @@ class SpecialExport extends SpecialPage { } else { $history = WikiExporter::CURRENT; } - + if( $page != '' ) $this->doExport = true; } - + if( !$wgExportAllowHistory ) { // Override $history = WikiExporter::CURRENT; } - + $list_authors = $wgRequest->getCheck( 'listauthors' ); if ( !$this->curonly || !$wgExportAllowListContributors ) $list_authors = false ; - + if ( $this->doExport ) { $wgOut->disable(); // Cancel output buffering and gzipping if set // This should provide safer streaming for pages with history wfResetOutputBuffers(); - header( "Content-type: application/xml; charset=utf-8" ); + $wgRequest->response()->header( "Content-type: application/xml; charset=utf-8" ); if( $wgRequest->getCheck( 'wpDownload' ) ) { // Provide a sane filename suggestion $filename = urlencode( $wgSitename . '-' . wfTimestampNow() . '.xml' ); @@ -143,22 +143,22 @@ class SpecialExport extends SpecialPage { $this->doExport( $page, $history, $list_authors ); return; } - + $wgOut->addWikiMsg( 'exporttext' ); - + $form = Xml::openElement( 'form', array( 'method' => 'post', - 'action' => $this->getTitle()->getLocalUrl( 'action=submit' ) ) ); - $form .= Xml::inputLabel( wfMsg( 'export-addcattext' ) , 'catname', 'catname', 40 ) . ' '; + 'action' => $this->getTitle()->getLocalUrl( 'action=submit' ) ) ); + $form .= Xml::inputLabel( wfMsg( 'export-addcattext' ) , 'catname', 'catname', 40 ) . ' '; $form .= Xml::submitButton( wfMsg( 'export-addcat' ), array( 'name' => 'addcat' ) ) . '
'; - + if ( $wgExportFromNamespaces ) { - $form .= Xml::namespaceSelector( $nsindex, null, 'nsindex', wfMsg( 'export-addnstext' ) ) . ' '; + $form .= Xml::namespaceSelector( $nsindex, null, 'nsindex', wfMsg( 'export-addnstext' ) ) . ' '; $form .= Xml::submitButton( wfMsg( 'export-addns' ), array( 'name' => 'addns' ) ) . '
'; } - + $form .= Xml::element( 'textarea', array( 'name' => 'pages', 'cols' => 40, 'rows' => 10 ), $page, false ); $form .= '
'; - + if( $wgExportAllowHistory ) { $form .= Xml::checkLabel( wfMsg( 'exportcuronly' ), 'curonly', 'curonly', true ) . '
'; } else { @@ -171,45 +171,60 @@ class SpecialExport extends SpecialPage { // Enable this when we can do something useful exporting/importing image information. :) //$form .= Xml::checkLabel( wfMsg( 'export-images' ), 'images', 'wpExportImages', false ) . '
'; $form .= Xml::checkLabel( wfMsg( 'export-download' ), 'wpDownload', 'wpDownload', true ) . '
'; - + $form .= Xml::submitButton( wfMsg( 'export-submit' ), array( 'accesskey' => 's' ) ); $form .= Xml::closeElement( 'form' ); $wgOut->addHTML( $form ); } - + private function userCanOverrideExportDepth() { - global $wgUser; + global $wgUser; return $wgUser->isAllowed( 'override-export-depth' ); } - + /** * Do the actual page exporting - * @param string $page User input on what page(s) to export - * @param mixed $history one of the WikiExporter history export constants + * + * @param $page String: user input on what page(s) to export + * @param $history Mixed: one of the WikiExporter history export constants + * @param $list_authors Boolean: Whether to add distinct author list (when + * not returning full history) */ private function doExport( $page, $history, $list_authors ) { global $wgExportMaxHistory; - - /* Split up the input and look up linked pages */ - $inputPages = array_filter( explode( "\n", $page ), array( $this, 'filterPage' ) ); - $pageSet = array_flip( $inputPages ); - + + $pageSet = array(); // Inverted index of all pages to look up + + // Split up and normalize input + foreach( explode( "\n", $page ) as $pageName ) { + $pageName = trim( $pageName ); + $title = Title::newFromText( $pageName ); + if( $title && $title->getInterwiki() == '' && $title->getText() !== '' ) { + // Only record each page once! + $pageSet[$title->getPrefixedText()] = true; + } + } + + // Set of original pages to pass on to further manipulation... + $inputPages = array_keys( $pageSet ); + + // Look up any linked pages if asked... if( $this->templates ) { $pageSet = $this->getTemplates( $inputPages, $pageSet ); } - + if( $linkDepth = $this->pageLinkDepth ) { $pageSet = $this->getPageLinks( $inputPages, $pageSet, $linkDepth ); } - + /* // Enable this when we can do something useful exporting/importing image information. :) if( $this->images ) ) { $pageSet = $this->getImages( $inputPages, $pageSet ); } */ - + $pages = array_keys( $pageSet ); // Normalize titles to the same format and remove dupes, see bug 17374 @@ -228,7 +243,7 @@ class SpecialExport extends SpecialPage { $lb = wfGetLBFactory()->newMainLB(); $db = $lb->getConnection( DB_SLAVE ); $buffer = WikiExporter::STREAM; - + // This might take a while... :D wfSuppressWarnings(); set_time_limit(0); @@ -254,28 +269,27 @@ class SpecialExport extends SpecialPage { $title = Title::newFromText( $page ); if( is_null( $title ) ) continue; #TODO: perhaps output an tag or something. if( !$title->userCanRead() ) continue; #TODO: perhaps output an tag or something. - + $exporter->pageByTitle( $title ); } - + $exporter->closeStream(); if( $lb ) { $lb->closeAll(); } } - - + private function getPagesFromCategory( $title ) { global $wgContLang; - + $name = $title->getDBkey(); - + $dbr = wfGetDB( DB_SLAVE ); $res = $dbr->select( array('page', 'categorylinks' ), array( 'page_namespace', 'page_title' ), array('cl_from=page_id', 'cl_to' => $name ), __METHOD__, array('LIMIT' => '5000')); - + $pages = array(); while ( $row = $dbr->fetchObject( $res ) ) { $n = $row->page_title; @@ -283,22 +297,22 @@ class SpecialExport extends SpecialPage { $ns = $wgContLang->getNsText( $row->page_namespace ); $n = $ns . ':' . $n; } - + $pages[] = $n; } $dbr->freeResult($res); - + return $pages; } - + private function getPagesFromNamespace( $nsindex ) { global $wgContLang; - + $dbr = wfGetDB( DB_SLAVE ); $res = $dbr->select( 'page', array('page_namespace', 'page_title'), array('page_namespace' => $nsindex), __METHOD__, array('LIMIT' => '5000') ); - + $pages = array(); while ( $row = $dbr->fetchObject( $res ) ) { $n = $row->page_title; @@ -306,13 +320,14 @@ class SpecialExport extends SpecialPage { $ns = $wgContLang->getNsText( $row->page_namespace ); $n = $ns . ':' . $n; } - + $pages[] = $n; } $dbr->freeResult($res); - + return $pages; } + /** * Expand a list of pages to include templates used in those pages. * @param $inputPages array, list of titles to look up @@ -325,7 +340,7 @@ class SpecialExport extends SpecialPage { array( 'tl_namespace AS namespace', 'tl_title AS title' ), array( 'page_id=tl_from' ) ); } - + /** * Validate link depth setting, if available. */ @@ -346,17 +361,18 @@ class SpecialExport extends SpecialPage { */ return intval( min( $depth, 5 ) ); } - + /** Expand a list of pages to include pages linked to from that page. */ private function getPageLinks( $inputPages, $pageSet, $depth ) { for( $depth=$depth; $depth>0; --$depth ) { $pageSet = $this->getLinks( $inputPages, $pageSet, 'pagelinks', array( 'pl_namespace AS namespace', 'pl_title AS title' ), array( 'page_id=pl_from' ) ); + $inputPages = array_keys( $pageSet ); } return $pageSet; } - + /** * Expand a list of pages to include images used in those pages. * @param $inputPages array, list of titles to look up @@ -369,7 +385,7 @@ class SpecialExport extends SpecialPage { array( NS_FILE . ' AS namespace', 'il_to AS title' ), array( 'page_id=il_from' ) ); } - + /** * Expand a list of pages to include items used in those pages. * @private @@ -380,7 +396,7 @@ class SpecialExport extends SpecialPage { $title = Title::newFromText( $page ); if( $title ) { $pageSet[$title->getPrefixedText()] = true; - /// @fixme May or may not be more efficient to batch these + /// @todo Fixme: May or may not be more efficient to batch these /// by namespace when given multiple input pages. $result = $dbr->select( array( 'page', $table ), @@ -388,7 +404,7 @@ class SpecialExport extends SpecialPage { array_merge( $join, array( 'page_namespace' => $title->getNamespace(), - 'page_title' => $title->getDBKey() ) ), + 'page_title' => $title->getDBkey() ) ), __METHOD__ ); foreach( $result as $row ) { $template = Title::makeTitle( $row->namespace, $row->title ); @@ -398,12 +414,5 @@ class SpecialExport extends SpecialPage { } return $pageSet; } - - /** - * Callback function to remove empty strings from the pages array. - */ - private function filterPage( $page ) { - return $page !== '' && $page !== null; - } }