Re: [Rd] URL checks
Instead of a separate file to store such a list, would it be an idea to add versions of the \href{}{} and \url{} markup commands that are skipped by the URL checks? Best, Wolfgang >-Original Message- >From: R-devel [mailto:r-devel-boun...@r-project.org] On Behalf Of Spencer >Graves >Sent: Friday, 08 January, 2021 13:04 >To: r-devel@r-project.org >Subject: Re: [Rd] URL checks > > I also would be pleased to be allowed to provide "a list of known >false-positive/exceptions" to the URL tests. I've been challenged >multiple times regarding URLs that worked fine when I checked them. We >should not be required to do a partial lobotomy to pass R CMD check ;-) > > Spencer Graves > >On 2021-01-07 09:53, Hugo Gruson wrote: >> >> I encountered the same issue today with https://astrostatistics.psu.edu/. >> >> This is a trust chain issue, as explained here: >> https://whatsmychaincert.com/?astrostatistics.psu.edu. >> >> I've worked for a couple of years on a project to increase HTTPS >> adoption on the web and we noticed that this type of error is very >> common, and that website maintainers are often unresponsive to requests >> to fix this issue. >> >> Therefore, I totally agree with Kirill that a list of known >> false-positive/exceptions would be a great addition to save time to both >> the CRAN team and package developers. >> >> Hugo >> >> On 07/01/2021 15:45, Kirill Müller via R-devel wrote: >>> One other failure mode: SSL certificates trusted by browsers that are >>> not installed on the check machine, e.g. the "GEANT Vereniging" >>> certificate from https://relational.fit.cvut.cz/ . >>> >>> K >>> >>> On 07.01.21 12:14, Kirill Müller via R-devel wrote: Hi The URL checks in R CMD check test all links in the README and vignettes for broken or redirected links. In many cases this improves documentation, I see problems with this approach which I have detailed below. I'm writing to this mailing list because I think the change needs to happen in R's check routines. I propose to introduce an "allow-list" for URLs, to reduce the burden on both CRAN and package maintainers. Comments are greatly appreciated. Best regards Kirill # Problems with the detection of broken/redirected URLs ## 301 should often be 307, how to change? Many web sites use a 301 redirection code that probably should be a 307. For example, https://www.oracle.com and https://www.oracle.com/ both redirect to https://www.oracle.com/index.html with a 301. I suspect the company still wants oracle.com to be recognized as the primary entry point of their web presence (to reserve the right to move the redirection to a different location later), I haven't checked with their PR department though. If that's true, the redirect probably should be a 307, which should be fixed by their IT department which I haven't contacted yet either. $ curl -i https://www.oracle.com HTTP/2 301 server: AkamaiGHost content-length: 0 location: https://www.oracle.com/index.html ... ## User agent detection twitter.com responds with a 400 error for requests without a user agent string hinting at an accepted browser. $ curl -i https://twitter.com/ HTTP/2 400 ... ...Please switch to a supported browser.. $ curl -s -i https://twitter.com/ -A "Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:84.0) Gecko/20100101 Firefox/84.0" | head -n 1 HTTP/2 200 # Impact While the latter problem *could* be fixed by supplying a browser-like user agent string, the former problem is virtually unfixable -- so many web sites should use 307 instead of 301 but don't. The above list is also incomplete -- think of unreliable links, HTTP links, other failure modes... This affects me as a package maintainer, I have the choice to either change the links to incorrect versions, or remove them altogether. I can also choose to explain each broken link to CRAN, this subjects the team to undue burden I think. Submitting a package with NOTEs delays the release for a package which I must release very soon to avoid having it pulled from CRAN, I'd rather not risk that -- hence I need to remove the link and put it back later. I'm aware of https://github.com/r-lib/urlchecker, this alleviates the problem but ultimately doesn't solve it. # Proposed solution ## Allow-list A file inst/URL that lists all URLs where failures are allowed -- possibly with a list of the HTTP codes accepted for that link. Example: https://oracle.com/ 301 https://twitter.com/drob/status/1224851726068527106 400 __ R-devel@r-project.org mailing list https://stat.ethz.ch/mailman/lis
Re: [Rd] URL checks
> Viechtbauer, Wolfgang (SP) > on Fri, 8 Jan 2021 13:50:14 + writes: > Instead of a separate file to store such a list, would it be an idea to add versions of the \href{}{} and \url{} markup commands that are skipped by the URL checks? > Best, > Wolfgang I think John Nash and you misunderstood -- or then I misunderstood -- the original proposal: I've been understanding that there should be a "central repository" of URL exceptions that is maintained by volunteers. And rather *not* that package authors should get ways to skip URL checking.. Martin >> -Original Message- >> From: R-devel [mailto:r-devel-boun...@r-project.org] On Behalf Of Spencer >> Graves >> Sent: Friday, 08 January, 2021 13:04 >> To: r-devel@r-project.org >> Subject: Re: [Rd] URL checks >> >> I also would be pleased to be allowed to provide "a list of known >> false-positive/exceptions" to the URL tests. I've been challenged >> multiple times regarding URLs that worked fine when I checked them. We >> should not be required to do a partial lobotomy to pass R CMD check ;-) >> >> Spencer Graves >> >> On 2021-01-07 09:53, Hugo Gruson wrote: >>> >>> I encountered the same issue today with https://astrostatistics.psu.edu/. >>> >>> This is a trust chain issue, as explained here: >>> https://whatsmychaincert.com/?astrostatistics.psu.edu. >>> >>> I've worked for a couple of years on a project to increase HTTPS >>> adoption on the web and we noticed that this type of error is very >>> common, and that website maintainers are often unresponsive to requests >>> to fix this issue. >>> >>> Therefore, I totally agree with Kirill that a list of known >>> false-positive/exceptions would be a great addition to save time to both >>> the CRAN team and package developers. >>> >>> Hugo >>> >>> On 07/01/2021 15:45, Kirill Müller via R-devel wrote: One other failure mode: SSL certificates trusted by browsers that are not installed on the check machine, e.g. the "GEANT Vereniging" certificate from https://relational.fit.cvut.cz/ . K On 07.01.21 12:14, Kirill Müller via R-devel wrote: > Hi > > The URL checks in R CMD check test all links in the README and > vignettes for broken or redirected links. In many cases this improves > documentation, I see problems with this approach which I have > detailed below. > > I'm writing to this mailing list because I think the change needs to > happen in R's check routines. I propose to introduce an "allow-list" > for URLs, to reduce the burden on both CRAN and package maintainers. > > Comments are greatly appreciated. > > Best regards > > Kirill > > # Problems with the detection of broken/redirected URLs > > ## 301 should often be 307, how to change? > > Many web sites use a 301 redirection code that probably should be a > 307. For example, https://www.oracle.com and https://www.oracle.com/ > both redirect to https://www.oracle.com/index.html with a 301. I > suspect the company still wants oracle.com to be recognized as the > primary entry point of their web presence (to reserve the right to > move the redirection to a different location later), I haven't > checked with their PR department though. If that's true, the redirect > probably should be a 307, which should be fixed by their IT > department which I haven't contacted yet either. > > $ curl -i https://www.oracle.com > HTTP/2 301 > server: AkamaiGHost > content-length: 0 > location: https://www.oracle.com/index.html > ... > > ## User agent detection > > twitter.com responds with a 400 error for requests without a user > agent string hinting at an accepted browser. > > $ curl -i https://twitter.com/ > HTTP/2 400 > ... > ...Please switch to a supported browser.. > > $ curl -s -i https://twitter.com/ -A "Mozilla/5.0 (X11; Ubuntu; Linux > x86_64; rv:84.0) Gecko/20100101 Firefox/84.0" | head -n 1 > HTTP/2 200 > > # Impact > > While the latter problem *could* be fixed by supplying a browser-like > user agent string, the former problem is virtually unfixable -- so > many web sites should use 307 instead of 301 but don't. The above > list is also incomplete -- think of unreliable links, HTTP links, > other failure modes... > > This affects me as a package maintainer, I have the choice to either > change the links to incorrect
Re: [Rd] URL checks
>> Viechtbauer, Wolfgang (SP) >> on Fri, 8 Jan 2021 13:50:14 + writes: > >> Instead of a separate file to store such a list, would it be an idea >to add versions of the \href{}{} and \url{} markup commands that are skipped >by the URL checks? >> Best, >> Wolfgang > >I think John Nash and you misunderstood -- or then I >misunderstood -- the original proposal: > >I've been understanding that there should be a "central repository" of URL >exceptions that is maintained by volunteers. > >And rather *not* that package authors should get ways to skip >URL checking.. > >Martin Hi Martin, Kirill suggested: "A file inst/URL that lists all URLs where failures are allowed -- possibly with a list of the HTTP codes accepted for that link." So, if it is a file in inst/, then this sounds to me like this is part of the package and not part of some central repository. Best, Wolfgang __ R-devel@r-project.org mailing list https://stat.ethz.ch/mailman/listinfo/r-devel
Re: [Rd] URL checks
> Viechtbauer, Wolfgang (SP) > on Mon, 11 Jan 2021 10:41:03 + writes: >>> Viechtbauer, Wolfgang (SP) >>> on Fri, 8 Jan 2021 13:50:14 + writes: >> >> > Instead of a separate file to store such a list, would it be an idea >> to add versions of the \href{}{} and \url{} markup commands that are skipped >> by the URL checks? >> > Best, >> > Wolfgang >> >> I think John Nash and you misunderstood -- or then I >> misunderstood -- the original proposal: >> >> I've been understanding that there should be a "central repository" of URL >> exceptions that is maintained by volunteers. >> >> And rather *not* that package authors should get ways to skip >> URL checking.. >> >> Martin > Hi Martin, > Kirill suggested: "A file inst/URL that lists all URLs where failures are allowed -- possibly with a list of the HTTP codes accepted for that link." > So, if it is a file in inst/, then this sounds to me like this is part of the package and not part of some central repository. > Best, > Wolfgang Dear Wolfgang, you are right and indeed it's *me* who misunderstood. But then I don't think it's a particularly good idea: From a CRAN point of view it is important that URLs in documents it hosts do not raise errors (*), hence the validity checking of URLs. So, CRAN (and other repository hosts) would need another option to still check all URLs .. and definitely would want to do that before accepting a package and also regularly do such checks on a per package basis in a way that it is reported as part of the CRAN checks of the respective package, right? So this will get envolved, ... and maybe it *is* good idea for a Google Summer of Code (GSoC) project ... well *if* it that is supervised by someone who's in close contact with CRAN or Bioc maintainer teams. Martin -- *) Such URL errors then lead to e-mails or other reports of web site checking engines reporting that you are hosting (too) (many) web pages with invalid links. __ R-devel@r-project.org mailing list https://stat.ethz.ch/mailman/listinfo/r-devel