X-Git-Url: http://git.cyclocoop.org/%22%20.%20generer_url_ecrire%28%22suivi_revisions%22%2C%22id_auteur=%24connecte%22%29%20.%20%22?a=blobdiff_plain;f=includes%2Fspecials%2FSpecialExport.php;h=3559ddee95acf9f0db2971d076850fb99ad129ce;hb=74a21f3bd1692dac958ddf3e09226a72b7bc65b7;hp=be58205f899f816bae21d7cb142d3ab5381bb7fd;hpb=0b6011413c8bf2c6d2f5bacff05b106dda3cb4f7;p=lhc%2Fweb%2Fwiklou.git
diff --git a/includes/specials/SpecialExport.php b/includes/specials/SpecialExport.php
index be58205f89..3559ddee95 100644
--- a/includes/specials/SpecialExport.php
+++ b/includes/specials/SpecialExport.php
@@ -22,40 +22,40 @@
*/
class SpecialExport extends SpecialPage {
-
+
private $curonly, $doExport, $pageLinkDepth, $templates;
private $images;
-
+
public function __construct() {
parent::__construct( 'Export' );
}
-
+
public function execute( $par ) {
global $wgOut, $wgRequest, $wgSitename, $wgExportAllowListContributors;
global $wgExportAllowHistory, $wgExportMaxHistory, $wgExportMaxLinkDepth;
global $wgExportFromNamespaces;
-
+
$this->setHeaders();
$this->outputHeader();
-
+
// Set some variables
$this->curonly = true;
$this->doExport = false;
$this->templates = $wgRequest->getCheck( 'templates' );
$this->images = $wgRequest->getCheck( 'images' ); // Doesn't do anything yet
$this->pageLinkDepth = $this->validateLinkDepth(
- $wgRequest->getIntOrNull( 'pagelink-depth' ) );
+ $wgRequest->getIntOrNull( 'pagelink-depth' ) );
$nsindex = '';
-
+
if ( $wgRequest->getCheck( 'addcat' ) ) {
$page = $wgRequest->getText( 'pages' );
$catname = $wgRequest->getText( 'catname' );
-
- if ( $catname !== '' && $catname !== NULL && $catname !== false ) {
+
+ if ( $catname !== '' && $catname !== null && $catname !== false ) {
$t = Title::makeTitleSafe( NS_MAIN, $catname );
if ( $t ) {
/**
- * @fixme This can lead to hitting memory limit for very large
+ * @todo Fixme: this can lead to hitting memory limit for very large
* categories. Ideally we would do the lookup synchronously
* during the export in a single query.
*/
@@ -67,14 +67,14 @@ class SpecialExport extends SpecialPage {
else if( $wgRequest->getCheck( 'addns' ) && $wgExportFromNamespaces ) {
$page = $wgRequest->getText( 'pages' );
$nsindex = $wgRequest->getText( 'nsindex', '' );
-
- if ( $nsindex !== '' && $nsindex !== null && $nsindex !== false ) {
+
+ if ( strval( $nsindex ) !== '' ) {
/**
- * Same implementation as above, so same @fixme
+ * Same implementation as above, so same @todo
*/
$nspages = $this->getPagesFromNamespace( $nsindex );
if ( $nspages ) $page .= "\n" . implode( "\n", $nspages );
- }
+ }
}
else if( $wgRequest->wasPosted() && $par == '' ) {
$page = $wgRequest->getText( 'pages' );
@@ -88,15 +88,15 @@ class SpecialExport extends SpecialPage {
$limit = $wgRequest->getInt( 'limit' );
$dir = $wgRequest->getVal( 'dir' );
$history = array(
- 'dir' => 'asc',
- 'offset' => false,
- 'limit' => $wgExportMaxHistory,
- );
+ 'dir' => 'asc',
+ 'offset' => false,
+ 'limit' => $wgExportMaxHistory,
+ );
$historyCheck = $wgRequest->getCheck( 'history' );
if ( $this->curonly ) {
$history = WikiExporter::CURRENT;
} elseif ( !$historyCheck ) {
- if ( $limit > 0 && $limit < $wgExportMaxHistory ) {
+ if ( $limit > 0 && ($wgExportMaxHistory == 0 || $limit < $wgExportMaxHistory ) ) {
$history['limit'] = $limit;
}
if ( !is_null( $offset ) ) {
@@ -106,7 +106,7 @@ class SpecialExport extends SpecialPage {
$history['dir'] = 'desc';
}
}
-
+
if( $page != '' ) $this->doExport = true;
} else {
// Default to current-only for GET requests
@@ -117,24 +117,24 @@ class SpecialExport extends SpecialPage {
} else {
$history = WikiExporter::CURRENT;
}
-
+
if( $page != '' ) $this->doExport = true;
}
-
+
if( !$wgExportAllowHistory ) {
// Override
$history = WikiExporter::CURRENT;
}
-
+
$list_authors = $wgRequest->getCheck( 'listauthors' );
if ( !$this->curonly || !$wgExportAllowListContributors ) $list_authors = false ;
-
+
if ( $this->doExport ) {
$wgOut->disable();
// Cancel output buffering and gzipping if set
// This should provide safer streaming for pages with history
wfResetOutputBuffers();
- header( "Content-type: application/xml; charset=utf-8" );
+ $wgRequest->response()->header( "Content-type: application/xml; charset=utf-8" );
if( $wgRequest->getCheck( 'wpDownload' ) ) {
// Provide a sane filename suggestion
$filename = urlencode( $wgSitename . '-' . wfTimestampNow() . '.xml' );
@@ -143,22 +143,22 @@ class SpecialExport extends SpecialPage {
$this->doExport( $page, $history, $list_authors );
return;
}
-
+
$wgOut->addWikiMsg( 'exporttext' );
-
+
$form = Xml::openElement( 'form', array( 'method' => 'post',
- 'action' => $this->getTitle()->getLocalUrl( 'action=submit' ) ) );
- $form .= Xml::inputLabel( wfMsg( 'export-addcattext' ) , 'catname', 'catname', 40 ) . ' ';
+ 'action' => $this->getTitle()->getLocalUrl( 'action=submit' ) ) );
+ $form .= Xml::inputLabel( wfMsg( 'export-addcattext' ) , 'catname', 'catname', 40 ) . ' ';
$form .= Xml::submitButton( wfMsg( 'export-addcat' ), array( 'name' => 'addcat' ) ) . '
';
-
+
if ( $wgExportFromNamespaces ) {
- $form .= Xml::namespaceSelector( $nsindex, null, 'nsindex', wfMsg( 'export-addnstext' ) ) . ' ';
+ $form .= Xml::namespaceSelector( $nsindex, null, 'nsindex', wfMsg( 'export-addnstext' ) ) . ' ';
$form .= Xml::submitButton( wfMsg( 'export-addns' ), array( 'name' => 'addns' ) ) . '
';
}
-
+
$form .= Xml::element( 'textarea', array( 'name' => 'pages', 'cols' => 40, 'rows' => 10 ), $page, false );
$form .= '
';
-
+
if( $wgExportAllowHistory ) {
$form .= Xml::checkLabel( wfMsg( 'exportcuronly' ), 'curonly', 'curonly', true ) . '
';
} else {
@@ -171,45 +171,60 @@ class SpecialExport extends SpecialPage {
// Enable this when we can do something useful exporting/importing image information. :)
//$form .= Xml::checkLabel( wfMsg( 'export-images' ), 'images', 'wpExportImages', false ) . '
';
$form .= Xml::checkLabel( wfMsg( 'export-download' ), 'wpDownload', 'wpDownload', true ) . '
';
-
+
$form .= Xml::submitButton( wfMsg( 'export-submit' ), array( 'accesskey' => 's' ) );
$form .= Xml::closeElement( 'form' );
$wgOut->addHTML( $form );
}
-
+
private function userCanOverrideExportDepth() {
- global $wgUser;
+ global $wgUser;
return $wgUser->isAllowed( 'override-export-depth' );
}
-
+
/**
* Do the actual page exporting
- * @param string $page User input on what page(s) to export
- * @param mixed $history one of the WikiExporter history export constants
+ *
+ * @param $page String: user input on what page(s) to export
+ * @param $history Mixed: one of the WikiExporter history export constants
+ * @param $list_authors Boolean: Whether to add distinct author list (when
+ * not returning full history)
*/
private function doExport( $page, $history, $list_authors ) {
global $wgExportMaxHistory;
-
- /* Split up the input and look up linked pages */
- $inputPages = array_filter( explode( "\n", $page ), array( $this, 'filterPage' ) );
- $pageSet = array_flip( $inputPages );
-
+
+ $pageSet = array(); // Inverted index of all pages to look up
+
+ // Split up and normalize input
+ foreach( explode( "\n", $page ) as $pageName ) {
+ $pageName = trim( $pageName );
+ $title = Title::newFromText( $pageName );
+ if( $title && $title->getInterwiki() == '' && $title->getText() !== '' ) {
+ // Only record each page once!
+ $pageSet[$title->getPrefixedText()] = true;
+ }
+ }
+
+ // Set of original pages to pass on to further manipulation...
+ $inputPages = array_keys( $pageSet );
+
+ // Look up any linked pages if asked...
if( $this->templates ) {
$pageSet = $this->getTemplates( $inputPages, $pageSet );
}
-
+
if( $linkDepth = $this->pageLinkDepth ) {
$pageSet = $this->getPageLinks( $inputPages, $pageSet, $linkDepth );
}
-
+
/*
// Enable this when we can do something useful exporting/importing image information. :)
if( $this->images ) ) {
$pageSet = $this->getImages( $inputPages, $pageSet );
}
*/
-
+
$pages = array_keys( $pageSet );
// Normalize titles to the same format and remove dupes, see bug 17374
@@ -228,7 +243,7 @@ class SpecialExport extends SpecialPage {
$lb = wfGetLBFactory()->newMainLB();
$db = $lb->getConnection( DB_SLAVE );
$buffer = WikiExporter::STREAM;
-
+
// This might take a while... :D
wfSuppressWarnings();
set_time_limit(0);
@@ -254,28 +269,27 @@ class SpecialExport extends SpecialPage {
$title = Title::newFromText( $page );
if( is_null( $title ) ) continue; #TODO: perhaps output an tag or something.
if( !$title->userCanRead() ) continue; #TODO: perhaps output an tag or something.
-
+
$exporter->pageByTitle( $title );
}
-
+
$exporter->closeStream();
if( $lb ) {
$lb->closeAll();
}
}
-
-
+
private function getPagesFromCategory( $title ) {
global $wgContLang;
-
+
$name = $title->getDBkey();
-
+
$dbr = wfGetDB( DB_SLAVE );
$res = $dbr->select( array('page', 'categorylinks' ),
array( 'page_namespace', 'page_title' ),
array('cl_from=page_id', 'cl_to' => $name ),
__METHOD__, array('LIMIT' => '5000'));
-
+
$pages = array();
while ( $row = $dbr->fetchObject( $res ) ) {
$n = $row->page_title;
@@ -283,22 +297,22 @@ class SpecialExport extends SpecialPage {
$ns = $wgContLang->getNsText( $row->page_namespace );
$n = $ns . ':' . $n;
}
-
+
$pages[] = $n;
}
$dbr->freeResult($res);
-
+
return $pages;
}
-
+
private function getPagesFromNamespace( $nsindex ) {
global $wgContLang;
-
+
$dbr = wfGetDB( DB_SLAVE );
$res = $dbr->select( 'page', array('page_namespace', 'page_title'),
array('page_namespace' => $nsindex),
__METHOD__, array('LIMIT' => '5000') );
-
+
$pages = array();
while ( $row = $dbr->fetchObject( $res ) ) {
$n = $row->page_title;
@@ -306,13 +320,14 @@ class SpecialExport extends SpecialPage {
$ns = $wgContLang->getNsText( $row->page_namespace );
$n = $ns . ':' . $n;
}
-
+
$pages[] = $n;
}
$dbr->freeResult($res);
-
+
return $pages;
}
+
/**
* Expand a list of pages to include templates used in those pages.
* @param $inputPages array, list of titles to look up
@@ -325,7 +340,7 @@ class SpecialExport extends SpecialPage {
array( 'tl_namespace AS namespace', 'tl_title AS title' ),
array( 'page_id=tl_from' ) );
}
-
+
/**
* Validate link depth setting, if available.
*/
@@ -346,17 +361,18 @@ class SpecialExport extends SpecialPage {
*/
return intval( min( $depth, 5 ) );
}
-
+
/** Expand a list of pages to include pages linked to from that page. */
private function getPageLinks( $inputPages, $pageSet, $depth ) {
for( $depth=$depth; $depth>0; --$depth ) {
$pageSet = $this->getLinks( $inputPages, $pageSet, 'pagelinks',
array( 'pl_namespace AS namespace', 'pl_title AS title' ),
array( 'page_id=pl_from' ) );
+ $inputPages = array_keys( $pageSet );
}
return $pageSet;
}
-
+
/**
* Expand a list of pages to include images used in those pages.
* @param $inputPages array, list of titles to look up
@@ -369,7 +385,7 @@ class SpecialExport extends SpecialPage {
array( NS_FILE . ' AS namespace', 'il_to AS title' ),
array( 'page_id=il_from' ) );
}
-
+
/**
* Expand a list of pages to include items used in those pages.
* @private
@@ -380,7 +396,7 @@ class SpecialExport extends SpecialPage {
$title = Title::newFromText( $page );
if( $title ) {
$pageSet[$title->getPrefixedText()] = true;
- /// @fixme May or may not be more efficient to batch these
+ /// @todo Fixme: May or may not be more efficient to batch these
/// by namespace when given multiple input pages.
$result = $dbr->select(
array( 'page', $table ),
@@ -388,7 +404,7 @@ class SpecialExport extends SpecialPage {
array_merge( $join,
array(
'page_namespace' => $title->getNamespace(),
- 'page_title' => $title->getDBKey() ) ),
+ 'page_title' => $title->getDBkey() ) ),
__METHOD__ );
foreach( $result as $row ) {
$template = Title::makeTitle( $row->namespace, $row->title );
@@ -398,12 +414,5 @@ class SpecialExport extends SpecialPage {
}
return $pageSet;
}
-
- /**
- * Callback function to remove empty strings from the pages array.
- */
- private function filterPage( $page ) {
- return $page !== '' && $page !== null;
- }
}