hi my life werey luckey m m like wold bat this posble my life On 5/4/17, [email protected] <[email protected]> wrote: > Send MediaWiki-l mailing list submissions to > [email protected] > > To subscribe or unsubscribe via the World Wide Web, visit > https://lists.wikimedia.org/mailman/listinfo/mediawiki-l > or, via email, send a message with subject or body 'help' to > [email protected] > > You can reach the person managing the list at > [email protected] > > When replying, please edit your Subject line so it is more specific > than "Re: Contents of MediaWiki-l digest..." > > > Today's Topics: > > 1. PluggableAuth and OAuth and login user experience (Scott Koranda) > 2. Re: Sitenotice for all pages except one (Jean Valjean) > 3. Re: Sitenotice for all pages except one (Sébastien Santoro) > 4. Load from a wget crawl (Rick Leir) > 5. Re: Load from a wget crawl (Brian Wolff) > 6. Re: Load from a wget crawl (G Rundlett) > 7. Re: Load from a wget crawl (Rick Leir) > > > ---------------------------------------------------------------------- > > Message: 1 > Date: Wed, 3 May 2017 15:10:03 -0500 > From: Scott Koranda <[email protected]> > To: [email protected] > Subject: [MediaWiki-l] PluggableAuth and OAuth and login user > experience > Message-ID: <[email protected]> > Content-Type: text/plain; charset=us-ascii > > Hi, > > I am using MediaWiki 1.28. > > I want users to authenticate using OpenID Connect so I have deployed the > PluggableAuth and OpenID Connect extensions and they are working well. > > I also want to provision accounts in my wiki from another system of record > using the API and so I have deployed the OAuth extension, created an > owner-only OAuth consumer, and have written a PHP client against the > API. It too is working well. > > The issue is that in order for the PHP client to leverage the API and > authenticate using OAuth AND for users to authenticate using OpenID > Connect I need to set > > $wgPluggableAuth_EnableLocalLogin = true; > > If I do not set that then the PHP client cannot authenticate using > OAuth. > > Have I missed something so that I would not have to enable local login > in order for the PHP client to use OAuth to authenticate and leverage > the API to provision accounts? > > If not, then I am satisfied with the solution I have except for the user > login experience. I want them to click "Log in" but not have to then see > the Special:UserLogin page. > > My thought is to replace that special page with one I create with a > custom extension that extends the SpecialUserLogin class as suggested > here: > > http://stackoverflow.com/questions/42776926/how-do-you-edit-the-html-for-mediawikis-specialuserlogin-page-in-mw-1-28 > > Is that the simplest and most elegant approach, or is there a cleaner > way to "hide" the Username/Password form from users and avoid them > having to click twice to start the OpenID Connect flow? > > Thanks, > > Scott K > > > > ------------------------------ > > Message: 2 > Date: Wed, 3 May 2017 17:22:06 -0400 > From: Jean Valjean <[email protected]> > To: MediaWiki announcements and site admin list > <[email protected]> > Subject: Re: [MediaWiki-l] Sitenotice for all pages except one > Message-ID: > <CAGLDgJN7=wuZw31QVT8J5Ju47bQ34fLS9h5rSXEfYwMwO5DQ=w...@mail.gmail.com> > Content-Type: text/plain; charset=UTF-8 > > Okay, this is the code I came up with: > > $wgAlreadyParsed = false; > function onParserBeforeStripSiteNotice( &$parser, &$text, &$strip_state ) { > global $wgAlreadyParsed; > if ( $wgAlreadyParsed ) { > return true; > } > $title = $parser->getTitle(); > if ( $title->getNamespace () !== 0 ) { > return true; > } > $wikiPage = WikiPage::factory( $title ); > $revision = $wikiPage->getRevision(); > if ( is_null ( $revision ) ) { > return true; > } > $content = $revision->getContent( Revision::RAW ); > $revisionText = ContentHandler::getContentText( $content ); > if ( $text !== $revisionText ) { > return true; > } > $wgAlreadyParsed = true; > $titleText = $title->getPrefixedText (); > $text = "{{siteNotice|$titleText}}" . $text; > } > $wgHooks['ParserBeforeStrip'][] = 'onParserBeforeStripSiteNotice'; > > Then in Template:SiteNotice I put: > > {{#ifeq: {{FULLPAGENAME}}|Main Page||<div class="plainlinks" > style="border:1px solid #a7d7f9; width:100%; font-size: 110%; text-align: > center; padding: 0.5ex; "><p>This page is an archive.</p></div>}} > > > On Wed, May 3, 2017 at 3:51 AM, Jean Valjean <[email protected]> > wrote: > >> How do I put up a sitenotice for all pages except one (e.g. Main Page)? I >> want the main page to have my current content, and all other pages to have >> a notice saying their content is just an archive. >> >> I notice that when I put {{FULLPAGENAME}} in MediaWiki:Sitenotice, it >> always says "MediaWiki:Sitenotice". >> >> Otherwise, I would use an #ifeq. >> > > > ------------------------------ > > Message: 3 > Date: Wed, 3 May 2017 23:40:57 +0200 > From: Sébastien Santoro <[email protected]> > To: MediaWiki announcements and site admin list > <[email protected]> > Subject: Re: [MediaWiki-l] Sitenotice for all pages except one > Message-ID: > <CAKg6iAETp60h=wwj4-wk8ut3jrtob-woemz998fzm465v1g...@mail.gmail.com> > Content-Type: text/plain; charset=UTF-8 > > Create an extension it will be easier ton maintain. > > Divide the code in several methods too, > https://jeroendedauw.github.io/slides/craftmanship/functions/#/1 > > On Wed, May 3, 2017 at 11:22 PM, Jean Valjean <[email protected]> > wrote: >> Okay, this is the code I came up with: >> >> $wgAlreadyParsed = false; >> function onParserBeforeStripSiteNotice( &$parser, &$text, &$strip_state ) >> { >> global $wgAlreadyParsed; >> if ( $wgAlreadyParsed ) { >> return true; >> } >> $title = $parser->getTitle(); >> if ( $title->getNamespace () !== 0 ) { >> return true; >> } >> $wikiPage = WikiPage::factory( $title ); >> $revision = $wikiPage->getRevision(); >> if ( is_null ( $revision ) ) { >> return true; >> } >> $content = $revision->getContent( Revision::RAW ); >> $revisionText = ContentHandler::getContentText( $content ); >> if ( $text !== $revisionText ) { >> return true; >> } >> $wgAlreadyParsed = true; >> $titleText = $title->getPrefixedText (); >> $text = "{{siteNotice|$titleText}}" . $text; >> } >> $wgHooks['ParserBeforeStrip'][] = 'onParserBeforeStripSiteNotice'; >> >> Then in Template:SiteNotice I put: >> >> {{#ifeq: {{FULLPAGENAME}}|Main Page||<div class="plainlinks" >> style="border:1px solid #a7d7f9; width:100%; font-size: 110%; text-align: >> center; padding: 0.5ex; "><p>This page is an archive.</p></div>}} >> >> >> On Wed, May 3, 2017 at 3:51 AM, Jean Valjean <[email protected]> >> wrote: >> >>> How do I put up a sitenotice for all pages except one (e.g. Main Page)? I >>> want the main page to have my current content, and all other pages to >>> have >>> a notice saying their content is just an archive. >>> >>> I notice that when I put {{FULLPAGENAME}} in MediaWiki:Sitenotice, it >>> always says "MediaWiki:Sitenotice". >>> >>> Otherwise, I would use an #ifeq. >>> >> _______________________________________________ >> MediaWiki-l mailing list >> To unsubscribe, go to: >> https://lists.wikimedia.org/mailman/listinfo/mediawiki-l > > > > -- > Sébastien Santoro aka Dereckson > http://www.dereckson.be/ > > > > ------------------------------ > > Message: 4 > Date: Wed, 03 May 2017 17:44:20 -0400 > From: Rick Leir <[email protected]> > To: [email protected] > Subject: [MediaWiki-l] Load from a wget crawl > Message-ID: <[email protected]> > Content-Type: text/plain; charset=utf-8 > > Hi all > Oops. I have a wget crawl of an old wikimedia site, and would like to load > it into a newly installed wikimedia. No easy way, I am almost sure. Please > tell me if I missed something. Cheers -- Rick > -- > Sorry for being brief. Alternate email is rickleir at yahoo dot com > > ------------------------------ > > Message: 5 > Date: Wed, 3 May 2017 19:16:32 -0400 > From: Brian Wolff <[email protected]> > To: MediaWiki announcements and site admin list > <[email protected]> > Subject: Re: [MediaWiki-l] Load from a wget crawl > Message-ID: > <ca+oo+dv_pgpb_vdmct0+nj3pq0dbeohgn+uxb39oc7pa-zr...@mail.gmail.com> > Content-Type: text/plain; charset=UTF-8 > > Yeah, that's basically impossible. > > Maybe you could try and use parsoid to convert the html into > wikitext-ish stuff (but it would be lossy as its not parsoid-html > that's being converted) and also would still involve some programming > to connect all the parts. > > -- > Brian > > On Wed, May 3, 2017 at 5:44 PM, Rick Leir <[email protected]> wrote: >> Hi all >> Oops. I have a wget crawl of an old wikimedia site, and would like to load >> it into a newly installed wikimedia. No easy way, I am almost sure. Please >> tell me if I missed something. Cheers -- Rick >> -- >> Sorry for being brief. Alternate email is rickleir at yahoo dot com >> _______________________________________________ >> MediaWiki-l mailing list >> To unsubscribe, go to: >> https://lists.wikimedia.org/mailman/listinfo/mediawiki-l > > > > ------------------------------ > > Message: 6 > Date: Wed, 3 May 2017 19:57:10 -0400 > From: G Rundlett <[email protected]> > To: MediaWiki announcements and site admin list > <[email protected]> > Subject: Re: [MediaWiki-l] Load from a wget crawl > Message-ID: > <CANaytcebYQXx8Zrvxzi-FKQCtkWn_FMd9=efzyuk8wws8fg...@mail.gmail.com> > Content-Type: text/plain; charset=UTF-8 > > You could try my html2wiki extension > > On May 3, 2017 7:17 PM, "Brian Wolff" <[email protected]> wrote: > >> Yeah, that's basically impossible. >> >> Maybe you could try and use parsoid to convert the html into >> wikitext-ish stuff (but it would be lossy as its not parsoid-html >> that's being converted) and also would still involve some programming >> to connect all the parts. >> >> -- >> Brian >> >> On Wed, May 3, 2017 at 5:44 PM, Rick Leir <[email protected]> wrote: >> > Hi all >> > Oops. I have a wget crawl of an old wikimedia site, and would like to >> load it into a newly installed wikimedia. No easy way, I am almost sure. >> Please tell me if I missed something. Cheers -- Rick >> > -- >> > Sorry for being brief. Alternate email is rickleir at yahoo dot com >> > _______________________________________________ >> > MediaWiki-l mailing list >> > To unsubscribe, go to: >> > https://lists.wikimedia.org/mailman/listinfo/mediawiki-l >> >> _______________________________________________ >> MediaWiki-l mailing list >> To unsubscribe, go to: >> https://lists.wikimedia.org/mailman/listinfo/mediawiki-l >> > > > ------------------------------ > > Message: 7 > Date: Wed, 3 May 2017 21:33:17 -0400 > From: Rick Leir <[email protected]> > To: [email protected] > Subject: Re: [MediaWiki-l] Load from a wget crawl > Message-ID: <[email protected]> > Content-Type: text/plain; charset=utf-8; format=flowed > > Thanks, that looks promising! > > > On 2017-05-03 07:57 PM, G Rundlett wrote: >> You could try my html2wiki extension >> >> On May 3, 2017 7:17 PM, "Brian Wolff" <[email protected]> wrote: >> >>> Yeah, that's basically impossible. >>> >>> Maybe you could try and use parsoid to convert the html into >>> wikitext-ish stuff (but it would be lossy as its not parsoid-html >>> that's being converted) and also would still involve some programming >>> to connect all the parts. >>> >>> -- >>> Brian >>> >>> On Wed, May 3, 2017 at 5:44 PM, Rick Leir <[email protected]> wrote: >>>> Hi all >>>> Oops. I have a wget crawl of an old wikimedia site, and would like to >>> load it into a newly installed wikimedia. No easy way, I am almost sure. >>> Please tell me if I missed something. Cheers -- Rick >>>> -- >>>> Sorry for being brief. Alternate email is rickleir at yahoo dot com >>>> _______________________________________________ >>>> MediaWiki-l mailing list >>>> To unsubscribe, go to: >>>> https://lists.wikimedia.org/mailman/listinfo/mediawiki-l >>> _______________________________________________ >>> MediaWiki-l mailing list >>> To unsubscribe, go to: >>> https://lists.wikimedia.org/mailman/listinfo/mediawiki-l >>> >> _______________________________________________ >> MediaWiki-l mailing list >> To unsubscribe, go to: >> https://lists.wikimedia.org/mailman/listinfo/mediawiki-l > > > > > ------------------------------ > > Subject: Digest Footer > > _______________________________________________ > MediaWiki-l mailing list > [email protected] > https://lists.wikimedia.org/mailman/listinfo/mediawiki-l > > > ------------------------------ > > End of MediaWiki-l Digest, Vol 164, Issue 3 > ******************************************* >
_______________________________________________ MediaWiki-l mailing list To unsubscribe, go to: https://lists.wikimedia.org/mailman/listinfo/mediawiki-l
