-
-
Notifications
You must be signed in to change notification settings - Fork 25.9k
[invalid] Dead link in latent dirichlet aloc #10275
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
sphinx provides Please provide a PR fixing this specific instance. |
My bad, I was watching old revision... abb43c1 |
scikit-learn.org is hosted on Github Pages, and that's not really suitable for archiving data.
That could be useful, the questions is how to run it using the infrastructure available for OSS projects. Done as a side project, this could be for instance a cron job on Travis CI for some repo (other than the scikit-learn one, where it's used already), then uploading the list of broken urls on some server (again different than scikit-learn.org). Note sure. |
Good to know. From what I saw, running multiple cron jobs with different tasks doesn't look easy as |
The easiest is probably to run |
For the record, you can have multiple builds with |
I meant that AFAIK it doesn't work as regular cron jobs would: one can't run two different taks with different periodicity. But I guess that's not a bit issue. |
it doesn't belong in a usual circle run. Linkcheck is a very slow process.
Even a separate bot running it quarterly and posting an issue with the
results would do fine
|
|
Not worth running on maintenance branch. Might be worth running on master,
except that a temporary web site outage and then recovery would send an
alert, which would suck. Basically, we only want it to tell us if there's
something broken for more than a week. Is there a nice way to do that?
|
I'm a bit ambivalent about sphinx's Checking for broken links in the documentation is a recurrent problem in OSS projects, and can be done with a few lines of scrapy (see e.g. here) independently of the way the docs were build, so instead of putting the effort into hacking the current CI setup (with all the constraints that is has, think about notifications etc) I think it could make sense to do this in a separate Github repo, check links with scrapy (or some other similar solution) and run it in Travis Cron. This way it could be applied to the different versions of scikit-learn docs without building them, scikit-learn-contrib projects or any other projects for that matter.. I can allocate some time for such a project. |
sounds like a good idea.
|
Is there bots / services that could do that? I would expect this to be mostly solvable statically without installing anything. |
Things like https://www.deadlinkchecker.com/?
…On 15 December 2017 at 07:26, Andreas Mueller ***@***.***> wrote:
Is there bots / services that could do that? I would expect this to be
mostly solvable statically without installing anything.
—
You are receiving this because you commented.
Reply to this email directly, view it on GitHub
<#10275 (comment)>,
or mute the thread
<https://github.com/notifications/unsubscribe-auth/AAEz6-MOjs1OlYN4bmH40jELeknXta6Wks5tAYR3gaJpZM4Q8A0r>
.
|
Uh oh!
There was an error while loading. Please reload this page.
Description
The link in documentation is dead here: https://github.com/scikit-learn/scikit-learn/blob/a24c8b46/sklearn/decomposition/online_lda.py#L260
I have a proposal to prevent these:
GET
them and check if the code is 200 or something. This would point out to dead links.The text was updated successfully, but these errors were encountered: