1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
|
# badlinks.py - plugin to list bad links
#
# Copyright (C) 1998, 1999 Albert Hopkins (marduk)
# Copyright (C) 2002 Mike W. Meyer
# Copyright (C) 2005, 2006, 2007, 2011, 2013 Arthur de Jong
#
# This program is free software; you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation; either version 2 of the License, or
# (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with this program; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
#
# The files produced as output from the software do not automatically fall
# under the copyright of the software, unless explicitly stated otherwise.
"""Present a listing of links that point to non-existent pages."""
__title__ = 'bad links'
__author__ = 'Arthur de Jong'
__outputfile__ = 'badlinks.html'
from sqlalchemy.orm import joinedload
from webcheck.db import Session, Link
from webcheck.output import render
def postprocess(crawler):
"""Add all bad links as pageproblems on pages where they are linked."""
session = Session()
# find all links with link problems
links = session.query(Link).filter(Link.linkproblems.any()).options(joinedload(Link.linkproblems))
# TODO: probably make it a nicer query over all linkproblems
for link in links:
# add a reference to the problem map
for problem in link.linkproblems:
for parent in link.parents:
parent.add_pageproblem('bad link: %s: %s' % (link.url, problem))
session.commit()
session.close()
def generate(crawler):
"""Present the list of bad links."""
session = Session()
links = session.query(Link).filter(Link.linkproblems.any())
links = links.order_by(Link.url).options(joinedload(Link.linkproblems))
render(__outputfile__, crawler=crawler, title=__title__,
links=links)
session.close()
|