753 lines
25 KiB
Python
753 lines
25 KiB
Python
#!/usr/bin/env python3
|
||
# -*- coding: utf-8 -*-
|
||
#
|
||
# redfam.py
|
||
#
|
||
# Copyright 2017 GOLDERWEB – Jonathan Golder <jonathan@golderweb.de>
|
||
#
|
||
# This program is free software; you can redistribute it and/or modify
|
||
# it under the terms of the GNU General Public License as published by
|
||
# the Free Software Foundation; either version 3 of the License, or
|
||
# (at your option) any later version.
|
||
#
|
||
# This program is distributed in the hope that it will be useful,
|
||
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||
# GNU General Public License for more details.
|
||
#
|
||
# You should have received a copy of the GNU General Public License
|
||
# along with this program; if not, write to the Free Software
|
||
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston,
|
||
# MA 02110-1301, USA.
|
||
#
|
||
#
|
||
"""
|
||
Provides classes for working with RedFams
|
||
"""
|
||
|
||
import hashlib
|
||
import locale
|
||
import re
|
||
from datetime import datetime
|
||
|
||
import mwparserfromhell as mwparser # noqa
|
||
import pywikibot # noqa
|
||
from pywikibot.tools import deprecated # noqa
|
||
|
||
import jogobot
|
||
from lib.mysqlred import MysqlRedFam, text
|
||
|
||
|
||
class RedFam( MysqlRedFam ):
|
||
"""
|
||
Basic class for RedFams, containing the basic data structure
|
||
"""
|
||
|
||
def __init__( self, articlesList, beginning, ending=None, redpageid=None,
|
||
status=None, famhash=None, heading=None ):
|
||
"""
|
||
Generates a new RedFam object
|
||
|
||
@param articlesList list List of articles
|
||
@param beginning datetime Beginning date
|
||
@param ending datetime Ending date
|
||
@param red_page_id int MW pageid of containing RedPage
|
||
@param status str Status of RedFam
|
||
@param fam_hash str SHA1 hash of articlesList
|
||
@param heading str Original heading of RedFam (Link)
|
||
"""
|
||
|
||
# Having pywikibot.Site() is a good idea most of the time
|
||
self.site = pywikibot.Site()
|
||
|
||
super().__init__(
|
||
articlesList=articlesList,
|
||
beginning=beginning,
|
||
ending=ending,
|
||
redpageid=redpageid,
|
||
famhash=famhash,
|
||
heading=heading,
|
||
status=status,
|
||
articlesStatus=None
|
||
)
|
||
|
||
def __repr__( self ):
|
||
"""
|
||
Returns repression str of RedFam object
|
||
|
||
@returns str repr() string
|
||
"""
|
||
|
||
__repr = "RedFam( " + \
|
||
"articlesList=" + repr( self.articlesList ) + \
|
||
", heading=" + repr( self.heading ) + \
|
||
", beginning=" + repr( self.beginning ) + \
|
||
", ending=" + repr( self.ending ) + \
|
||
", red_page_id=" + repr( self.redpageid ) + \
|
||
", status=" + repr( self.status ) + \
|
||
", fam_hash=" + repr( self.famhash ) + \
|
||
", articlesStatus=" + repr( self.articlesStatus ) + \
|
||
" )"
|
||
|
||
return __repr
|
||
|
||
@classmethod
|
||
def calc_famhash(cls, articlesList ):
|
||
"""
|
||
Calculates the SHA-1 hash for the articlesList of redundance family.
|
||
Since we don't need security SHA-1 is just fine.
|
||
|
||
@returns str String with the hexadecimal hash digest
|
||
"""
|
||
|
||
h = hashlib.sha1()
|
||
# Since articlesList attr of RedFam will have always 8 Members we
|
||
# need to fill up smaller lists (longers will be cropped below).
|
||
while len( articlesList) < 8:
|
||
articlesList.append(None)
|
||
|
||
h.update( str( articlesList[:8] ).encode('utf-8') )
|
||
|
||
return h.hexdigest()
|
||
|
||
@classmethod
|
||
def flush_db_cache( cls ):
|
||
"""
|
||
Calls flush method of Mysql Interface class
|
||
"""
|
||
cls.session.commit()
|
||
|
||
def article_add_status(self, status, index=None, title=None ):
|
||
"""
|
||
Adds a status specified by status, to article (identified by title
|
||
or index in articlesList) status set
|
||
|
||
@param status Statusstring to add
|
||
@type status str
|
||
@param index Add to article with index in articlesList
|
||
@type index int
|
||
@param title Add to article with title in articlesList
|
||
@type title str
|
||
"""
|
||
if title and not index:
|
||
index = self.articlesList.index( title )
|
||
|
||
if isinstance( index, int ) and index < len(self.articlesList):
|
||
self.articlesStatus[index].add(status)
|
||
else:
|
||
raise IndexError( "No index given or wrong format!")
|
||
|
||
def article_remove_status(self, status, index=None, title=None, weak=True):
|
||
"""
|
||
Removes a status specified by status, from article (identified by title
|
||
or index in articlesList) status set
|
||
If weak is set to False it will throw a KeyError when trying to
|
||
remove a status not set.
|
||
|
||
@param status Statusstring to add
|
||
@type status str
|
||
@param index Remove from article with index in articlesList
|
||
@type index int
|
||
@param title Remove from article with title in articlesList
|
||
@type title str
|
||
@param weak Change behavior on missing status
|
||
@type bool
|
||
"""
|
||
if title and not index:
|
||
index = self.articlesList.index( title )
|
||
|
||
if isinstance( index, int ) and index < len(self.articlesList):
|
||
if weak:
|
||
self.articlesStatus[index].discard(status)
|
||
else:
|
||
self.articlesStatus[index].remove(status)
|
||
else:
|
||
raise IndexError( "No index given or wrong format!")
|
||
|
||
def article_has_status(self, status, index=None, title=None ):
|
||
"""
|
||
Adds a status specified by status, to articles (identified by title
|
||
or index in articlesList) status set
|
||
|
||
@param status Statusstring to add
|
||
@type status str
|
||
@param index Check article with index in articlesList
|
||
@type index int
|
||
@param title Check article with title in articlesList
|
||
@type title str
|
||
"""
|
||
if title and not index:
|
||
index = self.articlesList.index( title )
|
||
|
||
if isinstance( index, int ) and index < len(self.articlesList):
|
||
if status in self.articlesStatus[index]:
|
||
return True
|
||
else:
|
||
return False
|
||
else:
|
||
raise IndexError( "No index given or wrong format!")
|
||
|
||
|
||
class RedFamParser( RedFam ):
|
||
"""
|
||
Provides an interface to RedFam for adding/updating redundance families
|
||
while parsig redundance pages
|
||
"""
|
||
|
||
# Define the timestamp format
|
||
__timestamp_format = jogobot.config['redundances']['timestamp_format']
|
||
|
||
# Define section heading re.pattern
|
||
__sectionhead_pat = re.compile( r"^(.*\[\[.+\]\].*\[\[.+\]\].*)" )
|
||
|
||
# Define timestamp re.pattern
|
||
__timestamp_pat = re.compile( jogobot.config['redundances']
|
||
['timestamp_regex'] )
|
||
|
||
# Textpattern for recognisation of done-notices
|
||
__done_notice = ":<small>Archivierung dieses Abschnittes \
|
||
wurde gewünscht von:"
|
||
__done_notice2 = "{{Erledigt|"
|
||
|
||
def __init__( self, articlesList, heading, redpage, redpagearchive,
|
||
beginning, ending=None ):
|
||
"""
|
||
Creates a RedFam object based on data collected while parsing red_pages
|
||
combined with possibly former known data from db
|
||
|
||
@param redfam_heading str Wikitext heading of section
|
||
@param redpage page Pywikibot.page object
|
||
@param redpagearchive bool Is red_page an archive
|
||
@param beginning datetime Timestamp of beginning
|
||
str as strptime parseable string
|
||
@param ending datetime Timestamp of ending
|
||
str strptime parseable string
|
||
"""
|
||
|
||
# Calculates the sha1 hash over self._articlesList to
|
||
# rediscover known redundance families
|
||
famhash = type(self).calc_famhash(articlesList)
|
||
|
||
# Set object attributes:
|
||
self.redpage = redpage
|
||
|
||
# Parse Timestamps
|
||
beginning = self.__datetime(beginning)
|
||
if ending:
|
||
ending = self.__datetime(ending)
|
||
|
||
super().__init__( articlesList,
|
||
beginning,
|
||
ending=ending,
|
||
redpageid=redpage.page._pageid,
|
||
famhash=famhash,
|
||
heading=heading )
|
||
|
||
# Check status changes
|
||
self.check_status()
|
||
|
||
self.session.add(self)
|
||
|
||
def update( self, articlesList, heading, redpage, redpagearchive,
|
||
beginning, ending=None ):
|
||
|
||
self.articlesList = articlesList
|
||
self.heading = heading
|
||
self.redpage = redpage
|
||
self.redpageid = redpage.pageid
|
||
|
||
self.add_beginning( beginning )
|
||
|
||
if ending:
|
||
self.add_ending( ending )
|
||
|
||
self._redpagearchive = redpagearchive
|
||
|
||
# Check status changes
|
||
self.check_status()
|
||
|
||
@classmethod
|
||
def heading_parser( cls, heading ):
|
||
"""
|
||
Parses given red_fam_heading string and saves articles list
|
||
|
||
@param heading Heading of RedFam-Section
|
||
@type heading wikicode or mwparser-parseable
|
||
"""
|
||
|
||
# Parse string heading with mwparse again everytime
|
||
# In some cases the given wikicode is broken due to syntax errors
|
||
# (Task FS#77)
|
||
heading = mwparser.parse( str( heading ) )
|
||
|
||
articlesList = []
|
||
for link in heading.ifilter_wikilinks():
|
||
article = str( link.title )
|
||
|
||
# Split in title and anchor part
|
||
article = article.split("#", 1)
|
||
# Replace underscores in title with spaces
|
||
article[0] = article[0].replace("_", " ")
|
||
|
||
if len(article) > 1:
|
||
# other way round, replace spaces with underscores in anchors
|
||
article[1] = article[1].replace(" ", "_")
|
||
|
||
# Rejoin title and anchor
|
||
article = "#".join(article)
|
||
|
||
# Add to list
|
||
articlesList.append(article)
|
||
|
||
return articlesList
|
||
|
||
def add_beginning( self, beginning ):
|
||
"""
|
||
Adds the beginning date of a redundance diskussion to the object
|
||
|
||
@param datetime datetime Beginning date
|
||
"""
|
||
|
||
self.beginning = self.__datetime( beginning )
|
||
|
||
def add_ending( self, ending ):
|
||
"""
|
||
Adds the ending date of a redundance diskussion to the object.
|
||
|
||
@param datetime datetime Ending date
|
||
"""
|
||
|
||
self.ending = self.__datetime( ending )
|
||
|
||
def __datetime( self, timestamp ):
|
||
"""
|
||
Decides wether given timestamp is a parseable string or a
|
||
datetime object and returns a datetime object in both cases
|
||
|
||
@param datetime timestamp Datetime object
|
||
str timestamp Parseable string with timestamp
|
||
|
||
@returns datetime Datetime object
|
||
"""
|
||
|
||
# Make sure locale is set to 'de_DE.UTF-8' to prevent problems
|
||
# with wrong month abreviations in strptime
|
||
locale.setlocale(locale.LC_ALL, 'de_DE.UTF-8')
|
||
|
||
if( isinstance( timestamp, datetime ) ):
|
||
return timestamp
|
||
else:
|
||
result = datetime.strptime( timestamp,
|
||
type( self ).__timestamp_format )
|
||
return result
|
||
|
||
def check_status( self ):
|
||
"""
|
||
Handles detection of correct status
|
||
There are three possible stati:
|
||
- 0 Discussion running --> no ending, page is not an archive
|
||
- 1 Discussion over --> ending present, page is not an archive
|
||
- 2 Discussion archived --> ending (normaly) present, page is archive
|
||
- 3 and greater status was set by worker script, do not change it
|
||
"""
|
||
|
||
# No ending, discussion is running:
|
||
# Sometimes archived discussions also have no detectable ending
|
||
if not self.ending and not self.redpage.archive:
|
||
self.status.add("open")
|
||
else:
|
||
self.status.remove("open")
|
||
if not self.redpage.archive:
|
||
self.status.add("done")
|
||
else:
|
||
self.status.remove("done")
|
||
self.status.remove("open")
|
||
self.status.add("archived")
|
||
|
||
@classmethod
|
||
def is_section_redfam_cb( cls, heading ):
|
||
"""
|
||
Used as callback for wikicode.get_sections in redpage.parse to
|
||
select sections which are redfams
|
||
"""
|
||
# Because of strange behavior in some cases, parse heading again
|
||
# (Task FS#77)
|
||
heading = mwparser.parse( str( heading ) )
|
||
|
||
# Make sure we have min. two wikilinks in heading to assume a redfam
|
||
if len( heading.filter_wikilinks() ) >= 2:
|
||
return True
|
||
else:
|
||
return False
|
||
|
||
@classmethod
|
||
def parser( cls, text, redpage, isarchive=False ):
|
||
"""
|
||
Handles parsing of redfam section
|
||
|
||
@param text Text of RedFam-Section
|
||
@type text wikicode or mwparser-parseable
|
||
"""
|
||
|
||
# Parse heading with mwparse if needed
|
||
if not isinstance( text, mwparser.wikicode.Wikicode ):
|
||
text = mwparser.parse( text )
|
||
|
||
# Extract heading text
|
||
heading = next( text.ifilter_headings() ).title.strip()
|
||
|
||
# Extract beginnig and maybe ending
|
||
(beginning, ending) = RedFamParser.extract_dates( text, isarchive )
|
||
|
||
# Missing beginning (Task: FS#76)
|
||
# Use first day of month of reddisc
|
||
if not beginning:
|
||
match = re.search(
|
||
jogobot.config["redundances"]["reddiscs_onlyinclude_re"],
|
||
redpage.page.title() )
|
||
|
||
if match:
|
||
beginning = datetime.strptime(
|
||
"01. {month} {year}".format(
|
||
month=match.group(1), year=match.group(2)),
|
||
"%d. %B %Y" )
|
||
articlesList = RedFamParser.heading_parser( heading )
|
||
famhash = RedFamParser.calc_famhash( articlesList )
|
||
|
||
# Check for existing objects in DB first in current redpage
|
||
redfam = redpage.redfams.get(famhash)
|
||
|
||
with RedFamParser.session.no_autoflush:
|
||
if not redfam:
|
||
# Otherwise in db table
|
||
redfam = RedFamParser.session.query(RedFamParser).filter(
|
||
RedFamParser.famhash == famhash ).one_or_none()
|
||
|
||
if redfam:
|
||
# Existing redfams need to be updated
|
||
redfam.update( articlesList, str(heading), redpage, isarchive,
|
||
beginning, ending )
|
||
|
||
else:
|
||
# Create the RedFam object
|
||
redfam = RedFamParser( articlesList, str(heading),
|
||
redpage, isarchive, beginning, ending )
|
||
|
||
# Add redfam to redpage object
|
||
redpage.redfams.set( redfam )
|
||
|
||
@classmethod
|
||
def extract_dates( cls, text, isarchive=False ):
|
||
"""
|
||
Returns tuple of the first and maybe last timestamp of a section.
|
||
Last timestamp is only returned if there is a done notice or param
|
||
*isarchiv* is set to 'True'
|
||
|
||
@param text Text to search in
|
||
@type line Any Type castable to str
|
||
@param isarchive If true skip searching done notice (on archivepages)
|
||
@type isarchive bool
|
||
|
||
@returns Timestamps, otherwise None
|
||
@returntype tuple of strs
|
||
"""
|
||
|
||
# Match all timestamps
|
||
matches = cls.__timestamp_pat.findall( str( text ) )
|
||
if matches:
|
||
|
||
# First one is beginning
|
||
# Since some timestamps are broken we need to reconstruct them
|
||
# by regex match groups
|
||
beginning = ( matches[0][0] + ", " + matches[0][1] + ". " +
|
||
matches[0][2] + ". " + matches[0][3] )
|
||
|
||
# Last one maybe is ending
|
||
# Done notice format 1
|
||
# Done notice format 2
|
||
# Or on archivepages
|
||
if ( cls.__done_notice in text or
|
||
cls.__done_notice2 in text or
|
||
isarchive ):
|
||
|
||
ending = ( matches[-1][0] + ", " + matches[-1][1] + ". " +
|
||
matches[-1][2] + ". " + matches[-1][3] )
|
||
|
||
else:
|
||
ending = None
|
||
# Missing dates (Task: FS#76)
|
||
else:
|
||
beginning = None
|
||
ending = None
|
||
|
||
return (beginning, ending)
|
||
|
||
|
||
class RedFamWorker( RedFam ):
|
||
"""
|
||
Handles working with redundance families stored in database
|
||
where discussion is finished
|
||
"""
|
||
def __init__( self ):
|
||
|
||
super().__init__()
|
||
|
||
# Make sure locale is set to 'de_DE.UTF-8' to prevent problems
|
||
# with wrong month abreviations in strptime
|
||
locale.setlocale(locale.LC_ALL, 'de_DE.UTF-8')
|
||
|
||
def article_generator(self, # noqa
|
||
filter_existing=None, filter_redirects=None,
|
||
exclude_article_status=[],
|
||
onlyinclude_article_status=[] ):
|
||
"""
|
||
Yields pywikibot pageobjects for articles belonging to this redfams
|
||
in a generator
|
||
self.
|
||
|
||
@param filter_existing Set to True to only get existing pages
|
||
set to False to only get nonexisting pages
|
||
unset/None results in not filtering
|
||
@type filter_existing bool/None
|
||
@param filter_redirects Set to True to get only noredirectpages,
|
||
set to False to get only redirectpages,
|
||
unset/None results in not filtering
|
||
@type filter_redirects bool/None
|
||
|
||
"""
|
||
|
||
# Helper to leave multidimensional loop
|
||
# https://docs.python.org/3/faq/design.html#why-is-there-no-goto
|
||
class Continue(Exception):
|
||
pass
|
||
|
||
class Break(Exception):
|
||
pass
|
||
|
||
# Iterate over articles in redfam
|
||
for article in self.articlesList:
|
||
|
||
# To be able to control outer loop from inside child loops
|
||
try:
|
||
|
||
# Not all list elements contain articles
|
||
if not article:
|
||
raise Break()
|
||
|
||
page = pywikibot.Page( pywikibot.Link(article),
|
||
pywikibot.Site() )
|
||
|
||
# Filter existing pages if requested with filter_existing=False
|
||
if page.exists():
|
||
self.article_remove_status( "deleted", title=article )
|
||
if filter_existing is False:
|
||
raise Continue()
|
||
# Filter non existing Pages if requested with
|
||
# filter_existing=True
|
||
else:
|
||
self.article_add_status( "deleted", title=article )
|
||
if filter_existing:
|
||
raise Continue()
|
||
|
||
# Filter redirects if requested with filter_redirects=True
|
||
if page.isRedirectPage():
|
||
self.article_add_status( "redirect", title=article )
|
||
if filter_redirects:
|
||
raise Continue()
|
||
# Filter noredirects if requested with filter_redirects=False
|
||
else:
|
||
self.article_remove_status("redirect", title=article )
|
||
if filter_redirects is False:
|
||
raise Continue()
|
||
|
||
# Exclude by article status
|
||
for status in exclude_article_status:
|
||
if self.article_has_status( status, title=article ):
|
||
raise Continue()
|
||
|
||
# Only include by article status
|
||
for status in onlyinclude_article_status:
|
||
if not self.article_has_status( status, title=article ):
|
||
raise Continue()
|
||
|
||
# Proxy loop control to outer loop
|
||
except Continue:
|
||
continue
|
||
except Break:
|
||
break
|
||
|
||
# Yield filtered pages
|
||
yield page
|
||
|
||
def update_status( self ):
|
||
"""
|
||
Sets status to 3 when worked on
|
||
"""
|
||
for article in self.articlesList:
|
||
if not article:
|
||
break
|
||
|
||
if self.article_has_status( "sav_err", title=article ):
|
||
self.status.add( "sav_err" )
|
||
return
|
||
elif self.article_has_status( "note_rej", title=article ):
|
||
self.status.add( "note_rej" )
|
||
return
|
||
|
||
elif not self.article_has_status("deleted", title=article ) and \
|
||
not self.article_has_status("redirect", title=article) and\
|
||
not self.article_has_status("marked", title=article):
|
||
return
|
||
|
||
self.status.remove("sav_err")
|
||
self.status.remove("note_rej")
|
||
self.status.add( "marked" )
|
||
|
||
def get_disc_link( self ):
|
||
"""
|
||
Constructs and returns the link to Redundancy discussion
|
||
|
||
@returns Link to diskussion
|
||
@rtype str
|
||
"""
|
||
|
||
# Expand templates using pwb site object
|
||
site = pywikibot.Site()
|
||
anchor_code = site.expand_text(self.heading.strip())
|
||
|
||
# Remove possibly embbeded files
|
||
anchor_code = re.sub( r"\[\[\w+:[^\|]+(?:\|.+){2,}\]\]", "",
|
||
anchor_code )
|
||
|
||
# Replace non-breaking-space by correct urlencoded value
|
||
anchor_code = anchor_code.replace( " ", ".C2.A0" )
|
||
|
||
# Use mwparser to strip and normalize
|
||
anchor_code = mwparser.parse( anchor_code ).strip_code()
|
||
|
||
# We try it without any more parsing as mw will do while parsing page
|
||
return ( self.redpage.pagetitle + "#" + anchor_code.strip() )
|
||
|
||
def generate_disc_notice_template( self ):
|
||
"""
|
||
Generates notice template to add on discussion Pages of Articles when
|
||
redundancy discussion is finished
|
||
|
||
@return Notice template to add on article disc
|
||
@rtype wikicode-node
|
||
"""
|
||
|
||
# Generate template boilerplate
|
||
template = mwparser.nodes.template.Template(
|
||
jogobot.config['redundances']['disc_notice_template_name'])
|
||
|
||
# Index of first article's param
|
||
param_cnt = 3
|
||
|
||
# Iterate over articles in redfam
|
||
for article in self.articlesList:
|
||
if not article:
|
||
break
|
||
# Make sure to only use 8 articles (max. param 10)
|
||
if param_cnt > 10:
|
||
break
|
||
|
||
# Add param for article
|
||
template.add( param_cnt, article, True )
|
||
|
||
param_cnt += 1
|
||
|
||
# Add begin
|
||
begin = self.beginning.strftime( "%B %Y" )
|
||
template.add( "Beginn", begin, True )
|
||
|
||
# Add end (if not same as begin)
|
||
end = self.ending.strftime( "%B %Y" )
|
||
if not end == begin:
|
||
template.add( "Ende", end, True )
|
||
|
||
# Add link to related reddisc
|
||
template.add( "Diskussion", self.get_disc_link(), True )
|
||
|
||
# Add signature and timestamp
|
||
# Not used atm
|
||
# template.add( 1, "-- ~~~~", True )
|
||
|
||
return template
|
||
|
||
@classmethod
|
||
def list_by_status( cls, status ):
|
||
"""
|
||
Lists red_fams stored in db by given status
|
||
"""
|
||
mysql = MysqlRedFam()
|
||
for fam in mysql.get_by_status( status ):
|
||
try:
|
||
print( cls( fam ) )
|
||
except RedFamHashError:
|
||
print(fam)
|
||
raise
|
||
|
||
@classmethod
|
||
def gen_by_status_and_ending( cls, status, ending ):
|
||
"""
|
||
Yield red_fams stored in db by given status which have an ending after
|
||
given one
|
||
"""
|
||
for redfam in RedFamWorker.session.query(RedFamWorker).filter(
|
||
# NOT WORKING WITH OBJECT NOTATION
|
||
# RedFamWorker._status.like('archived'),
|
||
# RedFamWorker._status.like("%{0:s}%".format(status)),
|
||
text("status LIKE '%archived%'"),
|
||
text("status NOT LIKE '%marked%'"),
|
||
RedFamWorker.ending >= ending ):
|
||
|
||
yield redfam
|
||
|
||
|
||
class RedFamError( Exception ):
|
||
"""
|
||
Base class for all Errors of RedFam-Module
|
||
"""
|
||
|
||
def __init__( self, message=None ):
|
||
"""
|
||
Handles Instantiation of RedFamError's
|
||
"""
|
||
if not message:
|
||
self.message = "An Error occured while executing a RedFam action"
|
||
else:
|
||
self.message = message
|
||
|
||
def __str__( self ):
|
||
"""
|
||
Output of error message
|
||
"""
|
||
|
||
return self.message
|
||
|
||
|
||
class RedFamHashError( RedFamError ):
|
||
"""
|
||
Raised when given RedFamHash does not match with calculated
|
||
"""
|
||
|
||
def __init__( self, givenHash, calculatedHash ):
|
||
|
||
message = "Given fam_hash ('{given}') does not match with \
|
||
calculated ('{calc}'".format( given=givenHash, calc=calculatedHash )
|
||
|
||
super().__init__( message )
|
||
|
||
|
||
class RedFamHeadingError ( RedFamError ):
|
||
"""
|
||
Raised when given RedFamHeading does not match __sectionhead_pat Regex
|
||
"""
|
||
def __init__( self, heading ):
|
||
|
||
message = "Error while trying to parse section heading. Given heading \
|
||
'{heading}' does not match RegEx".format( heading=heading )
|
||
|
||
super().__init__( message )
|