I would think the scripts lsarra mentioned would work best for MediaWiki. That said, there are a large number of 'archiving' tools that let you take copies of websites for historical preservation and backup:
e.g. https://github.com/pirate/ArchiveBox/wiki/Web-Archiving-Community#The-Master-Lists One noteworthy group - the 'Archive Team' https://www.archiveteam.org/ ~ Greg Greg Rundlett https://eQuality-Tech.com https://freephile.org On Sat, Mar 16, 2019 at 8:18 PM K. Peachey <[email protected]> wrote: > There is also MWOffliner which might work you, which is maintained by > the kiwix team iirc. > > On Sun, 17 Mar 2019 at 00:04, Isarra Yos <[email protected]> wrote: > > > > Not exactly the same thing, but there is a set of grabber scripts > > designed to get and import an entire wiki via the api, but these are > > currently mostly just mw maintenance and a few python scripts that only > > expect to be used with an actual second mediawiki instance. > > > > That being said, it probably would be a good idea to rearchitect them to > > be a bit less redundant with each other, and in so doing also make their > > usage a tad more flexible anyway, such as for creating dumps and backups > > instead. Whatever the case, at very least they might make a useful > > reference point. > > > > See: https://www.mediawiki.org/wiki/Manual:Grabbers > > > > -I > > > > On 15/03/2019 18:43, John wrote: > > > Unfortunately that doesn’t really work unless it’s a fairly small > wiki. If > > > it’s bigger the export request times out. > > > > > > What I was hoping for was a tool that utilizes either special export > or the > > > API to build a dump file using multiple requests. > > > > > > I can probably write something to do the same thing, but was hoping > for an > > > existing solution > > > > > > On Fri, Mar 15, 2019 at 1:12 PM Sterkebak <[email protected]> wrote: > > > > > >> XML export if enabled should do the trick. > > >> > > >> On Fri, 15 Mar 2019 at 18:04, John <[email protected]> wrote: > > >> > > >>> Are there any tools that allow you to create a dump of a mediawiki > > >> install > > >>> that doesn’t require direct database access? My primary focus is on > > >>> creating a backup of the wiki contents. > > >>> Thanks > > >>> _______________________________________________ > > >>> MediaWiki-l mailing list > > >>> To unsubscribe, go to: > > >>> https://lists.wikimedia.org/mailman/listinfo/mediawiki-l > > >>> > > >> -- > > >> Met vriendelijke groet, > > >> > > >> Huib Laurens > > >> _______________________________________________ > > >> MediaWiki-l mailing list > > >> To unsubscribe, go to: > > >> https://lists.wikimedia.org/mailman/listinfo/mediawiki-l > > >> > > > _______________________________________________ > > > MediaWiki-l mailing list > > > To unsubscribe, go to: > > > https://lists.wikimedia.org/mailman/listinfo/mediawiki-l > > > > > > > > _______________________________________________ > > MediaWiki-l mailing list > > To unsubscribe, go to: > > https://lists.wikimedia.org/mailman/listinfo/mediawiki-l > > _______________________________________________ > MediaWiki-l mailing list > To unsubscribe, go to: > https://lists.wikimedia.org/mailman/listinfo/mediawiki-l > _______________________________________________ MediaWiki-l mailing list To unsubscribe, go to: https://lists.wikimedia.org/mailman/listinfo/mediawiki-l
