diff --git a/noinfo/wallets.py b/noinfo/wallets.py
deleted file mode 100755
index b5891a6d6..000000000
--- a/noinfo/wallets.py
+++ /dev/null
@@ -1,302 +0,0 @@
-#!/usr/bin/env python 
-import sys, os, operator, urllib, json, re, time
-from datetime import datetime
-
-# 2017 originally by Martin Green
-# 2018-08-27 edited Philip Sargent
-# 2019-03-02 extended to take command line argument of loser_dir and set mod time of index.html to be same as json file
-# 2019-12-17 extra output of links to troggle-generated trip data
-# 2019-12-31 bits to make website link-checker not barf so much. Added endswith() to .startswith() for notes, elev, plan filenames
-# 2020-01-21 Now we are using Windows10-WSL1, +links to expedition logbook on every generated page
-# 2020-03-15 Adding timestamp tovisible outputs, changing name of produced files to walletindex.html so that contents can be browsed
-
-loser_dir = "/home/expo/loser"
-#loser_dir = "/mnt/d/CUCC-Expo/Loser/"  # when running on Win10-WSL1
-#loser_dir = "/media/philip/SD-huge/CUCC-Expo/loser/"  # when running on xubuntu laptop 'barbie'
-
-if len(sys.argv) > 1 :
-	if sys.argv[1] != "":
-		loser_dir = sys.argv[1]
-
-dateTimeObj=datetime.now(tz=None)
-timestamp = dateTimeObj.strftime("%d-%b-%Y (%H:%M)")
-
-print "Loser repo (for svx files) is assumed to be in: " + loser_dir + "/"
-
-drawings_dir =  loser_dir[0:len(loser_dir)-5] + "drawings"
-print "Drawings repo (for drawings files) is assumed to be in: " + drawings_dir + "/"
-
-html_base = "<html><body>%(body)s</body></html>"
-html_year_index = html_base % {"body": "<H1>%(year)s surveys: wallets status</H1>\n<p>List of trips: <a href=\"http://expo.survex.com/expedition/%(year)s\">expedition/%(year)s</a> - troggle-processed .svx files and logbook entries on server</p>\nAs of %(timestamp)s\n<H2>Persons</H2>\n<UL>\n%(persons)s</UL>\n<H2>Wallets</H2>\n<table>%(wallets)s</table>\n<H2>Needing Scanning</H2>\n<UL>\n%(needing scanning)s</ul>\n<H2>Website (Guidebook description) needing updating\n</H2>\n<UL style=\"column-count: 3; \">\n%(website needing updating)s</ul>\n"}
-html_year_person = "<li><a href='%(person)s.html'>%(person)s</a><UL>\n%(complaints)s</ul></li>\n"
-html_year_wallet_entry = "<tr><td><a href='%(walletindex)s'>%(walletname)s %(cave)s %(name)s</a></td> <td>%(complaints)s</td></tr>\n"
-html_person_wallet_entry = "<li><a href='%(walletindex)s'>%(walletname)s</a> <UL>\n%(complaints)s</ul></li>\n"
-html_year_scanning_entry = "<li><a href='%(walletindex)s'>%(walletname)s %(cave)s %(name)s</a></li>\n"
-html_wallet_file_entry = "<li><a href='%(fileurl)s'>%(filename)s</a></li>\n"
-html_wallet_index = html_base % {"body": "<H1>%(title)s : %(cave)s : %(name)s</H1>\n<p>List of trips: <a href=\"http://expo.survex.com/expedition/%(year)s\">expedition/%(year)s</a> - troggle-processed .svx files and logbook entries on server</p>\n<p>Date: %(date)s</p><p>People: %(people)s</p>\n<p>Cave <a href='%(description)s'>Guidebook description</a> - %(description_needed)s \n<p>Survex file:<br>&nbsp;&nbsp;<br>&nbsp;&nbsp;Local: &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;<a href='file:///%(loser_dir)s/%(survex)s' download>file:///%(loser_dir)s/%(survex)s</a><br>&nbsp;Server: &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;<a href='http://expo.survex.com/survexfile/%(survex)s' download>http://expo.survex.com/survexfile/%(survex)s</a></p><a href='../walletindex.html'>Wallet index for this year</a><br/>Local location for ::loser:: repo specified on command line is  <a href='file:///%(loser_dir)s'>%(loser_dir)s</a>. </p>\n<H2>Issues</H2>\n%(complaints)s\n<H2>Files</H2>\n<UL>\n%(files)s</UL>\n"}
-html_survex_required = {True: "Survex ", False: ""}
-html_plan_scanned = {True: "", False: "Plan "}
-html_elev_scanned = {True: "", False: "Elev "}
-html_description_written = {True: "", False: "Desc "}
-html_qms_written = {True: "", False: "QMs "}
-html_status = {True: "Issues: ", False: ""}
-html_person = html_base % {"body": "<H1>%(person)s</H1>\n<p>List of trips: <a href=\"http://expo.survex.com/expedition/%(year)s\">expedition/%(year)s</a> - troggle-processed .svx files and logbook entries on server</p>\n<H2>Outstanding Wallets</H2>\nAs of %(timestamp)s\n<UL>\n%(wallets)s</UL>"}
-html_complaint_items = "<li>%(count)i %(complaint)s</li>"
-html_items = "<li>%s</li>"
-
-
-
-blank_json = {
- "cave": "", 
- "date": "", 
- "description url": "/caves", 
- "description written": False, 
- "electronic survey": False, 
- "elev drawn": False, 
- "elev not required": False, 
- "name": "", 
- "people": [
-  "Unknown"
- ], 
- "plan drawn": False, 
- "plan not required": False, 
- "qms written": False, 
- "survex file": "", 
- "survex not required": False, 
- "website updated": False}
-
-
-#need to use wallets as a dict/tuple (id,cave,name) - not sure how.
-wallets = []
-wallets_needing_scanning = set()
-website_needing_updating = set()
-people = {}
-
-#use dir this file is in to get current year
-path,year = os.path.split(os.path.dirname(os.path.realpath(__file__)))
-
-print "Year: " + year
-
-for item in sorted(os.listdir(".")):
-    if os.path.isdir(item):
-        files = []
-        for f in os.listdir(os.path.join(".", item)):
-            if f not in ["contents.json", "contents.json~","walletindex.html"] and os.path.isfile(os.path.join(".", item, f)):
-                files.append(f)
-        contents_path = os.path.join(".", item, "contents.json")
-        #print "Trying to read file %s" % (contents_path) 
-        if not os.path.isfile(contents_path):
-            print "Creating file %s from template" % (contents_path) 
-            json_file = open(contents_path, "w")
-            json.dump(blank_json, json_file, sort_keys=True, indent = 1) 
-            json_file.close()
-        #print "Reading file %s" % (contents_path)
-        json_file = open(contents_path)
-        try:
-            data = json.load(json_file)
-        except:
-            print "FAILURE parsing JSON file %s" % (contents_path)
-            # Python bug: https://github.com/ShinNoNoir/twitterwebsearch/issues/12
-        if not data["people"]:
-            data["people"]=["NOBODY"]
-        json_file.close()
-        write_required = False
-        try:
-            wallet, cave, name = re.match("(\d\d\d\d#\d\d)-(.*) (.*)", item).groups()
-        except:
-            wallet, cave, name = "", "", ""
-        #print data
-        for k, v in blank_json.items():
-            if not data.has_key(k):
-                if k == "cave":
-                    data[k] = cave
-                elif k == "name":    
-                    data[k] = name
-                else:    
-                    data[k] = v
-                write_required = True
-        #print write_required
-        if write_required:
-            print "Writing file %s" % (contents_path) 
-            json_file = open(contents_path, "w")
-            json.dump(data, json_file, indent = 1)
-            json_file.close() 			
-						
-	# Get modification time of contents.json
-	# print("json last modified: %s" % time.ctime(os.path.getmtime(contents_path)))
-	json_mtime = os.path.getmtime(contents_path)
-
-        #make wallet descriptions
-        
-        #Survex
-        survex_required = (data["survex not required"] and data["survex file"] == "") or \
-                          not (not data["survex not required"] and os.path.isfile(os.path.join(loser_dir, data["survex file"])))
-        survex_complaint = ""
-        if data["survex not required"] and data["survex file"] != "":
-            survex_complaint = "Survex is not required and yet there is a survex file!"    
-        if not data["survex not required"] and data["survex file"] == "":
-            survex_complaint = "A survex file is required, but has not been specified!"    
-        if not data["survex not required"] and not os.path.isfile(os.path.join(loser_dir, data["survex file"])):
-            survex_complaint = "The specified survex file (%s) does not exist here!" % os.path.join(loser_dir, data["survex file"])
-        complaints = []
-        person_complaints = []
-        if survex_required:
-            complaints.append(survex_complaint)
-            person_complaints.append(survex_complaint)
-        
-        #Notes
-        notes_scanned = reduce(operator.or_, [f.startswith("note") for f in files], False)
-        notes_scanned = reduce(operator.or_, [f.endswith("note") for f in files], notes_scanned)
-        if not notes_scanned:
-            complaints.append("The notes needs scanning (no noteNN.jpg or XXnote.jpg file found)") 
-            wallets_needing_scanning.add(item)
-        
-        #Plan drawing required
-        plan_scanned = reduce(operator.or_, [f.startswith("plan") for f in files], False)
-        plan_scanned = reduce(operator.or_, [f.endswith("plan") for f in files], plan_scanned)
-        plan_drawing_required = not (plan_scanned or data["plan drawn"])
-        if plan_drawing_required:
-            complaints.append("The plan needs drawing (no planNN.jpg or XXplan.jpg file found)") 
-            person_complaints.append(" plan(s) needs drawing (no planNN.jpg or XXplan.jpg file found)") 
-        if not plan_drawing_required and not plan_scanned:
-            complaints.append("The plan needs  <em>scanning</em> (no planNN.jpg or XXplan.jpg file found)") 
-            wallets_needing_scanning.add(item)
-
-        
-        #Elev drawing required
-        elev_scanned = reduce(operator.or_, [f.startswith("elev") for f in files], False)
-        elev_scanned = reduce(operator.or_, [f.endswith("elev") for f in files], elev_scanned)
-        elev_drawing_required = not (elev_scanned or data["elev drawn"])
-        if elev_drawing_required:
-            complaints.append("The elev needs drawing (no elevNN.jpg or XXelev.jpg file found)") 
-            person_complaints.append(" elev(s) needs drawing (no elevNN.jpg or XXelev.jpg file found)") 
-        if not elev_drawing_required and not elev_scanned:
-            complaints.append("The elev needs <em>scanning</em> (no elevNN.jpg or XXelev.jpg file found)")
-            wallets_needing_scanning.add(item)
-
-        #Description
-        if not data["description written"]: 
-           complaints.append("The description needs writing") 
-           person_complaints.append(" description(s) needs writing")
-	description_needed = "A description is indicated as being needed, so may need adding into this cave page."
- 	
-
-        #QMS
-        if not data["qms written"]:
-            complaints.append("The QMs needs writing") 
-            person_complaints.append(" set(s) of QMs needs writing") 
-
-        #Website
-        if not data["website updated"]:
-            complaints.append("The website is marked as needing updating (using the guidebook description)")
-            website_needing_updating.add(item) 
-
-        #Electronic Surveys
-        if not data["electronic survey"]:
-            complaints.append("Tunnel / Therion drawing files need drawing") 
-
-        if data["survex file"]:
-            survex_description = data["survex file"]
-        else:
-            survex_description = "Not specified"
-
-        wallet_index_file = open(os.path.join(item, "walletindex.html"), "w")
-        wallet_index_file.write(html_wallet_index % {"title": item, "year": year,
-                                                     "cave": data["cave"],
-                                                     "name": data["name"],
-                                                     "date": data["date"],
-                                                     "people": reduce(operator.add, [" %s," % person for person in data["people"]], ""),
-                                                     "description": "http://expo.survex.com"+data["description url"],
-																										 "description_needed": description_needed,
-																										 "loser_dir": loser_dir,
-																										 "loser_dirw": loser_dir[5].upper() + ':/' + loser_dir[7:],
-                                                     "survex": survex_description,
-                                                     "complaints": reduce(operator.add, ["<p>" + complaint + "</p>" for complaint in complaints], ""),
-                                                     "files": reduce(operator.add, 
-                                                                     [html_wallet_file_entry % {"fileurl": urllib.quote(f), 
-                                                                                                "filename": f} 
-                                                                      for f
-                                                                      in files],
-                                                                     "")})
-        wallet_index_file.close()
-        wallets.append((item,  data["cave"],  data["name"], survex_required, plan_scanned, elev_scanned, data["description written"], data["qms written"]))
-	# Set modification time to be the same as that of contents.json
-	index_file = item+"/walletindex.html"
-	os.utime(index_file, ( json_mtime,json_mtime))
-
-				
-        #People
-
-        for person in data["people"]:
-            # delete all person.html as we are recreating all the ones that matter and old ones have old data
-            if os.path.isfile(person + ".html"):
-                os.remove(person + ".html")
-        if person_complaints:
-            for person in data["people"]:
-                if not people.has_key(person):
-                    people[person] = []
-                people[person].append((item, person_complaints))
-            
-
-wallets.sort()
-website_needing_updating = list(website_needing_updating)
-website_needing_updating.sort()
-wallets_needing_scanning = list(wallets_needing_scanning)
-wallets_needing_scanning.sort()
-
-person_summary = []
-for person, person_wallets in people.items():
-    complaints = reduce(operator.add, [complaints for wallet, complaints in person_wallets], [])
-    complaints_summary = []
-    for complaint in set(complaints):
-        complaints_summary.append((complaint, complaints.count(complaint)))
-    person_summary.append((person, complaints_summary))
-        
-person_summary = dict(person_summary)
-
-year_index_file = open("walletindex.html", "w")
-year_index_file.write(html_year_index % {"year": year, "timestamp": timestamp, "persons": reduce(operator.add,  [html_year_person % {"person": person,
-                                                                                               "complaints": reduce(operator.add,
-                                                                                                                    [html_complaint_items % {"complaint": complaint, 
-                                                                                                                                             "count": count}
-                                                                                                                     for complaint, count
-                                                                                                                     in complaints],
-                                                                                                                    "")} 
-                                                            for person, complaints
-                                                            in person_summary.items()], ""),
-                                         "needing scanning": reduce(operator.add,  [html_year_scanning_entry % {"walletname": wallet, 
-                                                                                      "cave": cave,
-                                                                                      "name": name,
-                                                                                      "walletindex": urllib.quote(wallet) + "/walletindex.html"} 
-                                                            for (wallet)
-                                                            in wallets_needing_scanning], ""),
-                                         "website needing updating": reduce(operator.add,  [html_year_scanning_entry % {"walletname": wallet,
-                                                                                      "cave": cave,
-                                                                                      "name": name,
-                                                                                      "walletindex": urllib.quote(wallet) + "/walletindex.html"} 
-                                                            for (wallet)
-                                                            in website_needing_updating], ""),
-                                         "wallets": reduce(operator.add, 
-                                                           [html_year_wallet_entry % {"walletname": wallet,
-                                                                                      "cave": cave,
-                                                                                      "name": name,
-                                                                                      "walletindex": urllib.quote(wallet) + "/walletindex.html", 
-                                                                                      "complaints": html_status[survex_required or not plan_scanned or not elev_scanned or description_written] + html_survex_required[survex_required] + html_plan_scanned[plan_scanned] + html_elev_scanned[elev_scanned] + html_description_written[description_written] + html_qms_written[qms_written] } 
-                                                            for (wallet, cave, name, survex_required, plan_scanned, elev_scanned, description_written, qms_written) 
-                                                            in wallets])})
-year_index_file.close()
-
-for person, item_complaint_list in people.items():
-    person_file = open(person + ".html", "w")
-    person_file.write(html_person % {"person": person, "year": year, "timestamp": timestamp,
-                                     "wallets": reduce(operator.add, [html_person_wallet_entry % {"walletname": wallet, 
-                                                                                                "walletindex": urllib.quote(wallet) + "/walletindex.html", 
-                                                                                                "complaints": reduce(operator.add, 
-                                                                                                                     [html_items % complaint 
-                                                                                                                      for complaint 
-                                                                                                                      in complaints], 
-           "") } 
-                                                                      for wallet, complaints 
-                                                                      in item_complaint_list], "")
-                                     })
-    person_file.close()
diff --git a/noinfo/walletscripts/wallets.py b/noinfo/walletscripts/wallets.py
index 94deb79d3..d51d1c056 100644
--- a/noinfo/walletscripts/wallets.py
+++ b/noinfo/walletscripts/wallets.py
@@ -1,47 +1,57 @@
 #!/usr/bin/env python 
 import sys, os, operator, urllib, json, re, time
+from datetime import datetime
 
 # 2017 originally by Martin Green
 # 2018-08-27 edited Philip Sargent
-# 2019-03-02 extended to take command line argument of loser_dir and set mod time of index.html to be sane as json file
+# 2019-03-02 extended to take command line argument of loser_dir and set mod time of index.html to be same as json file
 # 2019-12-17 extra output of links to troggle-generated trip data
-# 2019-12-31 bits to make website link-checker not barf so much. .startswith() replaced by .find() for notes, elev, plan filenames
+# 2019-12-31 bits to make website link-checker not barf so much. Added endswith() to .startswith() for notes, elev, plan filenames
+# 2020-01-21 Now we are using Windows10-WSL1, +links to expedition logbook on every generated page
+# 2020-03-15 Adding timestamp to visible outputs, changing name of produced files to walletindex.html so that contents can be browsed
+# 2020-03-15 Added "ignore" to the <year>#00 folder containing scraps - then removed as we do
+# want it to appear in the reports under "UNKNOWN"
 
-loser_dir = "/home/expo/loser/"
-#loser_dir = "/mnt/d/CUCC-Expo/loser/"  # when running on Win10/bash
+loser_dir = "/home/expo/loser"
+#loser_dir = "/mnt/d/CUCC-Expo/Loser/"  # when running on Win10-WSL1
 #loser_dir = "/media/philip/SD-huge/CUCC-Expo/loser/"  # when running on xubuntu laptop 'barbie'
 
 if len(sys.argv) > 1 :
 	if sys.argv[1] != "":
 		loser_dir = sys.argv[1]
 
-print "Loser repo (for svx files) is assumed to be in: " + loser_dir
+dateTimeObj=datetime.now(tz=None)
+timestamp = dateTimeObj.strftime("%d-%b-%Y (%H:%M)")
+
+print "Loser repo (for svx files) is assumed to be in: " + loser_dir + "/"
 
 drawings_dir =  loser_dir[0:len(loser_dir)-5] + "drawings"
-print "Drawings repo (for drawings files) is assumed to be in: " + drawings_dir
+print "Drawings repo (for drawings files) is assumed to be in: " + drawings_dir + "/"
 
 html_base = "<html><body>%(body)s</body></html>"
-html_year_index = html_base % {"body": "<H1>%(year)s surveys: wallets status</H1>\n<p>List of trips: <a href=\"http://expo.survex.com/expedition/%(year)s\">expedition/%(year)s</a> - troggle-processed .svx files and logbook entries on server</p>\n<H2>Persons</H2>\n<UL>\n%(persons)s</UL>\n<H2>Wallets</H2>\n<table>%(wallets)s</table>\n<H2>Needing Scanning</H2>\n<UL>\n%(needing scanning)s</ul>\n<H2>Website (Guidebook description) needing updating\n</H2>\n<UL>\n%(website needing updating)s</ul>\n"}
+html_year_index = html_base % {"body": "<H1>%(year)s surveys: wallets status</H1>\n<p>List of trips: <a href=\"http://expo.survex.com/expedition/%(year)s\">expedition/%(year)s</a> - troggle-processed .svx files and logbook entries on server</p>\nAs of %(timestamp)s\n<H2>Persons</H2>\n<UL>\n%(persons)s</UL>\n<H2>Wallets</H2>\n<table>%(wallets)s</table>\n<H2>Needing Scanning</H2>\n<UL>\n%(needing scanning)s</ul>\n<H2>Website (Guidebook description) needing updating\n</H2>\n<UL style=\"column-count: 3; \">\n%(website needing updating)s</ul>\n"}
 html_year_person = "<li><a href='%(person)s.html'>%(person)s</a><UL>\n%(complaints)s</ul></li>\n"
 html_year_wallet_entry = "<tr><td><a href='%(walletindex)s'>%(walletname)s %(cave)s %(name)s</a></td> <td>%(complaints)s</td></tr>\n"
 html_person_wallet_entry = "<li><a href='%(walletindex)s'>%(walletname)s</a> <UL>\n%(complaints)s</ul></li>\n"
 html_year_scanning_entry = "<li><a href='%(walletindex)s'>%(walletname)s %(cave)s %(name)s</a></li>\n"
 html_wallet_file_entry = "<li><a href='%(fileurl)s'>%(filename)s</a></li>\n"
-html_wallet_index = html_base % {"body": "<H1>%(title)s : %(cave)s : %(name)s</H1>\n<p>Date: %(date)s</p><p>People: %(people)s</p>\n<p>Cave <a href='%(description)s'>Guidebook description</a> - %(description_needed)s \n<p>Survex file:<br>&nbsp;&nbsp;Local (Windows): <a href='file://%(loser_dirw)s/%(survex)s' download>file://%(loser_dirw)s/%(survex)s</a><br>&nbsp;&nbsp;Local (Linux): <a href='file:///%(loser_dir)s/%(survex)s' download>file:///%(loser_dir)s/%(survex)s</a><br>&nbsp;&nbsp;Server: &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;<a href='http://expo.survex.com/survexfile/%(survex)s' download>http://expo.survex.com/survexfile/%(survex)s</a></p><a href='../index.html'>Wallet index for this year</a><br/>Local location for ::loser:: repo specified on command line is  <a href='file:///%(loser_dir)s'>%(loser_dir)s</a>. </p>\n<H2>Issues</H2>\n%(complaints)s\n<H2>Files</H2>\n<UL>\n%(files)s</UL>\n"}
+html_wallet_index = html_base % {"body": "<H1>%(title)s : %(cave)s : %(name)s</H1>\n<p>List of trips: <a href=\"http://expo.survex.com/expedition/%(year)s\">expedition/%(year)s</a> - troggle-processed .svx files and logbook entries on server</p>\n<p>Date: %(date)s</p><p>People: %(people)s</p>\n<p>Cave <a href='%(description)s'>Guidebook description</a> - %(description_needed)s \n<p>Survex file:<br>&nbsp;&nbsp;<br>&nbsp;&nbsp;Local: &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;<a href='file:///%(loser_dir)s/%(survex)s' download>file:///%(loser_dir)s/%(survex)s</a><br>&nbsp;Server: &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;<a href='http://expo.survex.com/survexfile/%(survex)s' download>http://expo.survex.com/survexfile/%(survex)s</a></p><a href='../walletindex.html'>Wallet index for this year</a><br/>Local location for ::loser:: repo specified on command line is  <a href='file:///%(loser_dir)s'>%(loser_dir)s</a>. </p>\n<H2>Issues</H2>\n%(complaints)s\n<H2>Files</H2>\n<UL>\n%(files)s</UL>\n"}
 html_survex_required = {True: "Survex ", False: ""}
 html_plan_scanned = {True: "", False: "Plan "}
 html_elev_scanned = {True: "", False: "Elev "}
 html_description_written = {True: "", False: "Desc "}
 html_qms_written = {True: "", False: "QMs "}
 html_status = {True: "Issues: ", False: ""}
-html_person = html_base % {"body": "<H1>%(person)s</H1><H2>Outstanding Wallets</H2><UL>\n%(wallets)s</UL>"}
+html_person = html_base % {"body": "<H1>%(person)s</H1>\n<p>List of trips: <a href=\"http://expo.survex.com/expedition/%(year)s\">expedition/%(year)s</a> - troggle-processed .svx files and logbook entries on server</p>\n<H2>Outstanding Wallets</H2>\nAs of %(timestamp)s\n<UL>\n%(wallets)s</UL>"}
 html_complaint_items = "<li>%(count)i %(complaint)s</li>"
 html_items = "<li>%s</li>"
 
+
+
 blank_json = {
  "cave": "", 
  "date": "", 
- "description url": "/1623/", 
+ "description url": "/caves", 
  "description written": False, 
  "electronic survey": False, 
  "elev drawn": False, 
@@ -70,26 +80,28 @@ path,year = os.path.split(os.path.dirname(os.path.realpath(__file__)))
 print "Year: " + year
 
 for item in sorted(os.listdir(".")):
-    if os.path.isdir(item):
+    if os.path.isdir(item) and item != year+"indexpages":
         files = []
         for f in os.listdir(os.path.join(".", item)):
-            if f not in ["contents.json", "contents.json~","index.html"] and os.path.isfile(os.path.join(".", item, f)):
+            if f not in ["contents.json", "contents.json~","walletindex.html"] and os.path.isfile(os.path.join(".", item, f)):
                 files.append(f)
         contents_path = os.path.join(".", item, "contents.json")
-#       print "Trying to read file %s" % (contents_path) 
+        #print "Trying to read file %s" % (contents_path) 
         if not os.path.isfile(contents_path):
             print "Creating file %s from template" % (contents_path) 
             json_file = open(contents_path, "w")
             json.dump(blank_json, json_file, sort_keys=True, indent = 1) 
             json_file.close()
-        # print "Reading file %s" % (contents_path) 
+        #print "Reading file %s" % (contents_path)
         json_file = open(contents_path)
-        #print json_file
-        data = json.load(json_file)
-	if not data["people"]:
-		data["people"]=["NOBODY"]
-
-
+        try:
+            data = json.load(json_file)
+        except:
+            print "FAILURE parsing JSON file %s" % (contents_path)
+            # Python bug: https://github.com/ShinNoNoir/twitterwebsearch/issues/12
+            raise
+        if not data["people"]:
+            data["people"]=["NOBODY"]
         json_file.close()
         write_required = False
         try:
@@ -191,8 +203,8 @@ for item in sorted(os.listdir(".")):
         else:
             survex_description = "Not specified"
 
-        wallet_index_file = open(os.path.join(item, "index.html"), "w")
-        wallet_index_file.write(html_wallet_index % {"title": item,
+        wallet_index_file = open(os.path.join(item, "walletindex.html"), "w")
+        wallet_index_file.write(html_wallet_index % {"title": item, "year": year,
                                                      "cave": data["cave"],
                                                      "name": data["name"],
                                                      "date": data["date"],
@@ -212,7 +224,7 @@ for item in sorted(os.listdir(".")):
         wallet_index_file.close()
         wallets.append((item,  data["cave"],  data["name"], survex_required, plan_scanned, elev_scanned, data["description written"], data["qms written"]))
 	# Set modification time to be the same as that of contents.json
-	index_file = item+"/index.html"
+	index_file = item+"/walletindex.html"
 	os.utime(index_file, ( json_mtime,json_mtime))
 
 				
@@ -245,8 +257,8 @@ for person, person_wallets in people.items():
         
 person_summary = dict(person_summary)
 
-year_index_file = open("index.html", "w")
-year_index_file.write(html_year_index % {"year": year, "persons": reduce(operator.add,  [html_year_person % {"person": person,
+year_index_file = open("walletindex.html", "w")
+year_index_file.write(html_year_index % {"year": year, "timestamp": timestamp, "persons": reduce(operator.add,  [html_year_person % {"person": person,
                                                                                                "complaints": reduce(operator.add,
                                                                                                                     [html_complaint_items % {"complaint": complaint, 
                                                                                                                                              "count": count}
@@ -258,20 +270,20 @@ year_index_file.write(html_year_index % {"year": year, "persons": reduce(operato
                                          "needing scanning": reduce(operator.add,  [html_year_scanning_entry % {"walletname": wallet, 
                                                                                       "cave": cave,
                                                                                       "name": name,
-                                                                                      "walletindex": urllib.quote(wallet) + "/index.html"} 
+                                                                                      "walletindex": urllib.quote(wallet) + "/walletindex.html"} 
                                                             for (wallet)
                                                             in wallets_needing_scanning], ""),
                                          "website needing updating": reduce(operator.add,  [html_year_scanning_entry % {"walletname": wallet,
                                                                                       "cave": cave,
                                                                                       "name": name,
-                                                                                      "walletindex": urllib.quote(wallet) + "/index.html"} 
+                                                                                      "walletindex": urllib.quote(wallet) + "/walletindex.html"} 
                                                             for (wallet)
                                                             in website_needing_updating], ""),
                                          "wallets": reduce(operator.add, 
                                                            [html_year_wallet_entry % {"walletname": wallet,
                                                                                       "cave": cave,
                                                                                       "name": name,
-                                                                                      "walletindex": urllib.quote(wallet) + "/index.html", 
+                                                                                      "walletindex": urllib.quote(wallet) + "/walletindex.html", 
                                                                                       "complaints": html_status[survex_required or not plan_scanned or not elev_scanned or description_written] + html_survex_required[survex_required] + html_plan_scanned[plan_scanned] + html_elev_scanned[elev_scanned] + html_description_written[description_written] + html_qms_written[qms_written] } 
                                                             for (wallet, cave, name, survex_required, plan_scanned, elev_scanned, description_written, qms_written) 
                                                             in wallets])})
@@ -279,9 +291,9 @@ year_index_file.close()
 
 for person, item_complaint_list in people.items():
     person_file = open(person + ".html", "w")
-    person_file.write(html_person % {"person": person, 
+    person_file.write(html_person % {"person": person, "year": year, "timestamp": timestamp,
                                      "wallets": reduce(operator.add, [html_person_wallet_entry % {"walletname": wallet, 
-                                                                                                "walletindex": urllib.quote(wallet) + "/index.html", 
+                                                                                                "walletindex": urllib.quote(wallet) + "/walletindex.html", 
                                                                                                 "complaints": reduce(operator.add, 
                                                                                                                      [html_items % complaint 
                                                                                                                       for complaint