[open-allure-ds] push by jg07...@gmail.com - Add parseForm to scriptParser on 2012-01-08 23:31 GMT

17 views
Skip to first unread message

open-al...@googlecode.com

unread,
Jan 8, 2012, 6:32:03 PM1/8/12
to open-al...@googlegroups.com
Revision: 88c0f48fee82
Author: John Graves
Date: Sun Jan 8 15:31:17 2012
Log: Add parseForm to scriptParser
http://code.google.com/p/open-allure-ds/source/detail?r=88c0f48fee82

Modified:
/wikitospeech/SlideSpeech.py
/wikitospeech/scriptParser.py

=======================================
--- /wikitospeech/SlideSpeech.py Tue Dec 6 20:16:21 2011
+++ /wikitospeech/SlideSpeech.py Sun Jan 8 15:31:17 2012
@@ -136,7 +136,8 @@
if len(seq.sequence[seq.onQuestion].pathToImageFiles)>0:
linkToShow = seq.sequence[seq.onQuestion].pathToImageFiles
+ linkToShow
else:
- linkToShow = "static/" + linkToShow
+ pass
+ #linkToShow = "static/" + linkToShow
if len(seq.sequence[seq.onQuestion].answers)>0:
return forms.showJPGSlideWithQuestion(linkToShow, \
seq.sequence[seq.onQuestion] )
=======================================
--- /wikitospeech/scriptParser.py Fri Nov 25 16:43:11 2011
+++ /wikitospeech/scriptParser.py Sun Jan 8 15:31:17 2012
@@ -42,8 +42,8 @@

def parseScript(name):
if name.startswith("http"):
- if name.find("etherpad")>0:
- sequence = parseEtherpad(name)
+ if name.find("wikitospeech.pagekite.me")>0:
+ sequence = parseForm(name)
elif name.find("titanpad")>0:
sequence = parseEtherpad(name)
else:
@@ -96,6 +96,46 @@
sequence = parseText(text)
return sequence

+def parseForm(name):
+ # download and extract script from wikitospeech.pagekite.me page, for
example:
+ # http://wikitospeech.pagekite.me?tag=Test
+ try:
+ proxy = os.environ["HTTP_PROXY"]
+ except:
+ proxy = ''
+ if proxy=="http://cache.aut.ac.nz:3128":
+ proxies = {'http': 'http://cache.aut.ac.nz:3128'}
+ urlOpen = urllib.urlopen( name , proxies=proxies )
+ else:
+ urlOpen = urllib.urlopen( name, proxies={} )
+
+ # extract text marked with <from> from wikitospeech page
+ soup = BeautifulSoup(urlOpen.read())
+ taggedPre = soup.form
+ img=[i['src'] for i in soup.form.findAll('img')]
+ #[u'http://dl.dropbox.com/u/12838403/20120109/img0.png',
u'http://dl.dropbox.com/u/12838403/20120109/img1.png']
+ text=[i.contents[0].strip() for i in soup.form.findAll('textarea')]
+ #[u'Notes on slide 1', u'Notes on slide 2']
+ urlText = []
+ for i in range(0,len(img)):
+ print i
+ urlText.append(img[i])
+ textLines = text[i].split("\n")
+ for line in textLines:
+ urlText.append(line)
+ urlText.append("\n")
+
+ f = open('debug.txt','w')
+ f.write("test run at " + strftime("%d %b %Y %H:%M", gmtime()) + "\n")
+ f.write(str(type(urlText))+ "\n")
+ f.write(str(len(urlText))+ "\n")
+ for l in urlText:
+ f.write(l.strip()+"\n")
+ f.close()
+
+ sequence = parseText(urlText)
+ return sequence
+
def parseHtml(name):
# download and extract script from wiki page, for example:
# http://dl.dropbox.com/u/12838403/dropbox.txt

Reply all
Reply to author
Forward
0 new messages