print '}'
-def data_main(d):
+def data_main(d, force=False):
d = {k: str(v) for k, v in dict(d).iteritems()}
structure_data(d)
d['matchdata'] = []
d['summarydawg'] = nodelists['Summary']
del(d['matchdata'])
crawl = crawler.Crawler()
- crawl.add_entry(d)
- crawl.write()
+ crawl.add_entry(d, force)
return crawl.list_names()
-def eq_pre(req, args):
- req.log_error('handler')
- req.content_type = 'text/html'
- req.send_http_header()
- data = (
- '<html>\n<head>\n'
- '\t<title>HyperFrontend RSS feed input</title>\n'
- '\t<script src="contextmenu_o.js"></script>\n</head>\n<body>\n'
- '\t<table>\n'
- '\t\t<tr><td>Venue: </td><td>\n'
- '\t\t\t<input type="text" name="venue" class="indat"></td></tr>\n'
- '\t\t<tr><td>Frequency: </td><td>\n'
- '\t\t\t<input type="text" name="freq" class="indat"></td></tr>\n'
- '\t\t<tr><td>Default location name: </td><td>\n'
- '\t\t\t<input type="text" name="dloc" class="indat"></td></tr>\n'
- '\t\t<tr><td>Adress: </td><td>\n'
- '\t\t\t<input type="text" name="adress" class="indat"></td></tr>\n'
- '\t\t<tr><td>Website: </td><td>\n'
- '\t\t\t<input type="text" name="website" class="indat"></td></tr>\n'
- '\t\t<tr><td>RSS URL: </td><td>\n'
- '\t\t\t<input type="text" name="url" class="indat"'
- 'value="{}"></td></tr>\n'
- '\t</table>\n'
- '\tSelecteer iets en klik de link<br />\n'
- '\t<button style="color:blue" onclick="javascript:f_waar()">'
- 'Waar</button>\n'
- '\t<button style="color:green" onclick="javascript:f_wat()">'
- 'Wat</button>\n'
- '\t<button style="color:red" onclick="javascript:f_wann_tijd()">'
- 'Tijd</button>\n'
- '\t<button style="color:darkred" onclick="javascript:f_wann_dat()">'
- 'Datum</button>\n').format(args['url'])
- req.write(data)
-
-
-def req_post(req):
- req.write('\t<button onclick="javascript:post_all()" method="post" '
- 'target="_blank">Submit</button>\n</body>\n</html>')
-
-
def feed2html(req, url, name):
url = urllib.unquote(url)
url = url if re.match('https?://', url) else 'http://{}'.format(url)