3 # helper program for getting information from yppedia
5 # This is part of ypp-sc-tools, a set of third-party tools for assisting
6 # players of Yohoho Puzzle Pirates.
8 # Copyright (C) 2009 Ian Jackson <ijackson@chiark.greenend.org.uk>
10 # This program is free software: you can redistribute it and/or modify
11 # it under the terms of the GNU General Public License as published by
12 # the Free Software Foundation, either version 3 of the License, or
13 # (at your option) any later version.
15 # This program is distributed in the hope that it will be useful,
16 # but WITHOUT ANY WARRANTY; without even the implied warranty of
17 # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
18 # GNU General Public License for more details.
20 # You should have received a copy of the GNU General Public License
21 # along with this program. If not, see <http://www.gnu.org/licenses/>.
23 # Yohoho and Puzzle Pirates are probably trademarks of Three Rings and
24 # are used without permission. This program is not endorsed or
25 # sponsored by Three Rings.
28 yppedia-ocean-scraper is part of ypp-sc-tools Copyright (C) 2009 Ian Jackson
29 This program comes with ABSOLUTELY NO WARRANTY; this is free software,
30 and you are welcome to redistribute it under certain conditions. For
31 details, read the top of the yppedia-ocean-scraper file.
35 signal.signal(signal.SIGINT, signal.SIG_DFL)
42 from optparse import OptionParser
43 from BeautifulSoup import BeautifulSoup
52 print >>sys.stderr, k,`v`
56 url = ('http://yppedia.puzzlepirates.com/%s_Ocean' %
57 urllib.quote(ocean,''))
59 dataf = urllib2.urlopen(url)
60 debug('fetched',dataf)
61 soup = BeautifulSoup(dataf)
64 title_arch_re = regexp.compile('(\\S.*\\S) Archipelago \\((\\S+)\\)$')
65 title_any_re = regexp.compile('(\\S.*\\S) \((\\S+)\\)$')
66 href_img_re = regexp.compile('\\.png$')
68 def title_arch_info(t):
69 # returns (arch,ocean)
71 if t is None: return (None,None)
72 m = title_arch_re.match(t)
73 if not m: return (None,None)
77 (a,o) = title_arch_info(t)
78 if o is None: return False
82 firstarch = soup.find('a', attrs = {'title': title_arch_ok})
85 def findall_title_arch_ok(t):
86 return t.findAll('a', attrs = {'title': title_arch_ok})
88 def is_archestable(u):
89 if u.name != 'table': return False
90 return len(findall_title_arch_ok(u)) > 1
92 archestable = firstarch.findParent('table', attrs={'border':'1'})
93 debug('at',archestable)
96 for row in archestable.findAll('tr',recursive=False):
97 archsoups += row.findAll('td',recursive=False)
101 return len(v.findAll(text = regexp.compile('.*Large'))) > 0
103 return u.findParent(is_island)
105 for arch in archsoups:
106 links = arch.findAll('a', href=True)
108 if not links: continue
109 (a,o) = title_arch_info(links[0]['title'])
110 debug('arch-ocean', (a,o))
112 assert(a not in arches)
114 for link in links[1:]:
116 if href_img_re.search(link['href']): continue
117 m = title_any_re.match(link['title'])
118 assert(m.group(2) == ocean)
120 debug('island', island)
131 for island in arches[a]:
139 '''usage: .../yppedia-ocean-scraper [--debug] OCEAN''')
141 ao('--debug', action='count', dest='debug', default=0,
142 help='enable debugging output')
144 (opts,args) = pa.parse_args()
146 print >>sys.stderr, copyright_info
147 pa.error('need an ocean argument')