You cannot select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
149 lines
5.5 KiB
Python
149 lines
5.5 KiB
Python
#! /usr/bin/env python
|
|
# -*- coding: utf-8 -*-
|
|
|
|
##############
|
|
# Edit the index sorting articles according to topic, section and issue
|
|
#####
|
|
|
|
#
|
|
import xml.etree.ElementTree as ET
|
|
import html5lib, urllib2, json, pprint, subprocess
|
|
from urllib import quote as quote
|
|
import re
|
|
|
|
sid = '1234'
|
|
useragent = "Mozilla/5.001 (windows; U; NT4.0; en-US; rv:1.0) Gecko/25250101"
|
|
endpoint = "http://pzwiki.wdka.nl/mw-mediadesign/api.php?format=json&"
|
|
allworks = {}
|
|
mainkeys = ['Thumbnail','Date','Creator']
|
|
|
|
def api_request(action, pagename): #get page: content, metadata, images, imageifnp
|
|
url = endpoint + action.format(pagename)
|
|
request = urllib2.urlopen(url)
|
|
jsonp = json.loads(request.read() )
|
|
json_dic= (jsonp.get('query').get('pages'))
|
|
page_id = json_dic.keys()[0]
|
|
page_content = json_dic.get(page_id)
|
|
print 'API URL:', url
|
|
return page_content
|
|
|
|
|
|
def api_page(pageid, info):
|
|
if info == 'content':
|
|
api_response = api_request('action=query&pageids={}&prop=revisions&rvprop=content', pageid)
|
|
page_content = ((api_response.get('revisions'))[0])['*']
|
|
elif info == 'metadata':
|
|
page_content = api_request('action=query&pageids={}&prop=info', pageid)
|
|
elif info == 'articleimgs':
|
|
page_content = api_request('action=query&pageids={}&prop=images', pageid)
|
|
elif info == 'imageinfo':
|
|
page_content = api_request('action=query&pageids={}&prop=imageinfo&iiprop=url&iiurlwidth=300', pageid) # iiurlwidht dermines with of thumbnail
|
|
return page_content
|
|
|
|
def api_page_content(pagename):
|
|
content = ((page.get('revisions'))[0])['*']
|
|
return content
|
|
# print json.dumps( revisions, sort_keys=True, indent=4) ## see response
|
|
|
|
|
|
|
|
def api_thumb_url(filename):
|
|
'''get thumbnail url of image'''
|
|
page_content_dict = api_page(filename, 'imageinfo')
|
|
if 'imageinfo' in page_content_dict.keys():
|
|
thumburl = ((page_content_dict.get('imageinfo'))[0].get('thumburl'))
|
|
return page_content_dict
|
|
|
|
|
|
def find_imgs(article):
|
|
'''get all internal images from published articles:
|
|
* query all images in a article
|
|
* for each image request a thumbnail
|
|
* get the thumburl
|
|
'''
|
|
page_content_dict = api_page(article, 'articleimgs')
|
|
if 'images' in page_content_dict.keys():
|
|
images_list = page_content_dict.get('images')
|
|
# print 'images_list', images_list
|
|
thumbs_list = []
|
|
for img in images_list: # all images in article
|
|
title = img.get('title')
|
|
thumburl_json = api_thumb_url(title)
|
|
if thumburl_json != None:
|
|
thumburl = (thumburl_json.get('imageinfo')[0]).get('thumburl')
|
|
thumbs_list.append(thumburl)
|
|
else:
|
|
thumburl = None
|
|
|
|
return thumbs_list
|
|
|
|
def parse_work_page(title, content):
|
|
content = content.encode('utf-8')
|
|
if re.match('\{\{\Graduation work', content):
|
|
work_dict = {}
|
|
work_dict['Title']=title
|
|
template, extra = (re.findall('\{\{Graduation work\n(.*?)\}\}(.*)', content, re.DOTALL))[0]
|
|
# template's key/value pair
|
|
keyval = re.findall('\|(.*?)\=(.*?\n)', template, re.DOTALL)
|
|
extra = ('Extra', extra)
|
|
keyval.append(extra)
|
|
for pair in keyval:
|
|
key = pair[0]
|
|
val = pair[1]
|
|
val = val.replace('\n','')
|
|
if 'Creator' in key:
|
|
val = val.replace(', ', '')
|
|
work_dict[key]=val
|
|
|
|
return work_dict
|
|
|
|
'''
|
|
TEMPLATE
|
|
|
|
|Description=
|
|
|Creator=
|
|
|Date=
|
|
|Thumbnail=
|
|
|Website=
|
|
|
|
Description=Based on her written thesis: The Web Cheated on Me, Marie is trying to figure out where her disappointment with the web comes from. She analyzed her webbrowser history for half a year to find out what kind of information she is looking up. Her graduation work is an audio installation based on this research.\n|Creator=Marie Wocher,\n|Date=2013\n|Thumbnail=4 FromHypertextToApplePie.jpg\n
|
|
'''
|
|
|
|
def api_category(category, year):
|
|
'''Finds all pages within category and returns a dictionary with info on those pages'''
|
|
category = category.replace(' ', '_')
|
|
if year:
|
|
api_url = endpoint + 'action=query&list=categorymembers&cmlimit=500&cmtitle=Category:{}&cmtitle=Category:{}'.format(category, year) #BUG: API only queries last cmtitle: YEAR
|
|
else:
|
|
api_url = endpoint + 'action=query&list=categorymembers&cmlimit=500&cmtitle=Category:{}'.format(category)
|
|
|
|
request = urllib2.urlopen(api_url)
|
|
jsonp = json.loads(request.read())
|
|
# dict_page = {}
|
|
for page in jsonp['query']['categorymembers']:
|
|
print 'Page:', page
|
|
title = ((page['title']).encode('utf-8') ).replace(" ", "_") #snakecase for page titles
|
|
pageid = page['pageid']
|
|
print 'Pageid:', pageid
|
|
## NOTE: instead of using page name to query page, use PAGE ID
|
|
article = api_page(pageid, 'content')
|
|
print 'Content:'
|
|
pprint.pprint(article)
|
|
print
|
|
work = parse_work_page(title, article) #
|
|
if work and set(mainkeys).issubset(work.keys()) and len([ work[key] for key in mainkeys if work[key] ])==3: # work must exist, have mainkeys as keys w/ values
|
|
allworks[pageid] = work
|
|
|
|
|
|
api_category('Graduation work', '2013')
|
|
#pprint.pprint(allworks)
|
|
|
|
# save json
|
|
json_allworks = open('md_allworks.json', 'w')
|
|
json.dump(allworks, json_allworks )
|
|
print "wrote json dictionary to:", 'md_allworks.json'
|
|
|
|
## TO DO
|
|
# How do handle work['Extra'] value?
|
|
# some tiles work['Extra'] contains: <gallery>, [[Pages]], text, etc
|