[pmwiki-users] Trouble with .pageindex when too much _new_ data to index (+ sqlite)

Patrick R. Michaud pmichaud at pobox.com
Wed Jan 28 15:47:38 CST 2015

On Wed, Jan 28, 2015 at 10:10:06PM +0100, ABClf wrote:
> Main issue encountered is how .pageindex is handling its indexation task.
> It sounds like it definitely stops working when the amount of _new_ data is
> too big. [...]

I don't think this is quite how it works.  If you're examining the code to
see how things work, it would help a lot if you could refer to functions by

Here are a couple of possibly helpful items, though.  The amount of time
PmWiki spends building .pageindex is controlled by the $PageIndexTime variable,
which defaults to 10 seconds.  This variable controls how long PmWiki spends
indexing pages after serving a page in response to a search; increasing the
value of this variable should allow more pages to be indexed during each

As far as I remember, PmWiki doesn't check to see if the amount of new 
data is acceptable to index.  PmWiki simply has a list of pages that it
knows aren't in the index, indexes as many as it can in $PageIndexTime 
seconds, and leaves the rest for a later request to index.  

If there's a single page that is too large for PmWiki to index because of 
PHP memory limits, I'm not sure there's a whole lot we can do about that 
short of increasing the memory limits.

I think it would not be difficult to create a custom action that does
nothing but pageindexing for a very large page or set of pages... but I
really need to know more precise details about what you're trying to do
before I can help further, such as specific page names and exactly what
they contain.


More information about the pmwiki-users mailing list