1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
|
<?php
/**
* See deferred.txt
* @package MediaWiki
*/
/**
*
* @package MediaWiki
*/
class SearchUpdate {
/* private */ var $mId = 0, $mNamespace, $mTitle, $mText;
/* private */ var $mTitleWords;
function SearchUpdate( $id, $title, $text = false ) {
$nt = Title::newFromText( $title );
if( $nt ) {
$this->mId = $id;
$this->mText = $text;
$this->mNamespace = $nt->getNamespace();
$this->mTitle = $nt->getText(); # Discard namespace
$this->mTitleWords = $this->mTextWords = array();
} else {
wfDebug( "SearchUpdate object created with invalid title '$title'\n" );
}
}
function doUpdate() {
global $wgContLang, $wgDisableSearchUpdate;
if( $wgDisableSearchUpdate || !$this->mId ) {
return false;
}
$fname = 'SearchUpdate::doUpdate';
wfProfileIn( $fname );
$search = SearchEngine::create();
$lc = $search->legalSearchChars() . '&#;';
if( $this->mText === false ) {
$search->updateTitle($this->mId,
Title::indexTitle( $this->mNamespace, $this->mTitle ));
wfProfileOut( $fname );
return;
}
# Language-specific strip/conversion
$text = $wgContLang->stripForSearch( $this->mText );
wfProfileIn( $fname.'-regexps' );
$text = preg_replace( "/<\\/?\\s*[A-Za-z][A-Za-z0-9]*\\s*([^>]*?)>/",
' ', strtolower( " " . $text /*$this->mText*/ . " " ) ); # Strip HTML markup
$text = preg_replace( "/(^|\\n)==\\s*([^\\n]+)\\s*==(\\s)/sD",
"\\1\\2 \\2 \\2\\3", $text ); # Emphasize headings
# Strip external URLs
$uc = "A-Za-z0-9_\\/:.,~%\\-+&;#?!=()@\\xA0-\\xFF";
$protos = "http|https|ftp|mailto|news|gopher";
$pat = "/(^|[^\\[])({$protos}):[{$uc}]+([^{$uc}]|$)/";
$text = preg_replace( $pat, "\\1 \\3", $text );
$p1 = "/([^\\[])\\[({$protos}):[{$uc}]+]/";
$p2 = "/([^\\[])\\[({$protos}):[{$uc}]+\\s+([^\\]]+)]/";
$text = preg_replace( $p1, "\\1 ", $text );
$text = preg_replace( $p2, "\\1 \\3 ", $text );
# Internal image links
$pat2 = "/\\[\\[image:([{$uc}]+)\\.(gif|png|jpg|jpeg)([^{$uc}])/i";
$text = preg_replace( $pat2, " \\1 \\3", $text );
$text = preg_replace( "/([^{$lc}])([{$lc}]+)]]([a-z]+)/",
"\\1\\2 \\2\\3", $text ); # Handle [[game]]s
# Strip all remaining non-search characters
$text = preg_replace( "/[^{$lc}]+/", " ", $text );
# Handle 's, s'
#
# $text = preg_replace( "/([{$lc}]+)'s /", "\\1 \\1's ", $text );
# $text = preg_replace( "/([{$lc}]+)s' /", "\\1s ", $text );
#
# These tail-anchored regexps are insanely slow. The worst case comes
# when Japanese or Chinese text (ie, no word spacing) is written on
# a wiki configured for Western UTF-8 mode. The Unicode characters are
# expanded to hex codes and the "words" are very long paragraph-length
# monstrosities. On a large page the above regexps may take over 20
# seconds *each* on a 1GHz-level processor.
#
# Following are reversed versions which are consistently fast
# (about 3 milliseconds on 1GHz-level processor).
#
$text = strrev( preg_replace( "/ s'([{$lc}]+)/", " s'\\1 \\1", strrev( $text ) ) );
$text = strrev( preg_replace( "/ 's([{$lc}]+)/", " s\\1", strrev( $text ) ) );
# Strip wiki '' and '''
$text = preg_replace( "/''[']*/", " ", $text );
wfProfileOut( "$fname-regexps" );
$search->update($this->mId, Title::indexTitle( $this->mNamespace, $this->mTitle ),
$text);
wfProfileOut( $fname );
}
}
/**
* Placeholder class
* @package MediaWiki
*/
class SearchUpdateMyISAM extends SearchUpdate {
# Inherits everything
}
?>
|