logbook typescripts, reports to funding agencies, copies of expo reports in magazines, instruction guides for how to run an expo, accounting protocols for subsidising transport, rope maintenance standards: over 900 pages in all. What you are reading now is part of this.<br><br>
<li>The <strong>file archive</strong>: the photo galleries, the copies of PDF reports, the scans of original survey notebooks, GPS tracks, maps, aerial photos, scanned images of our handwritten expo logbooks. There are 40 GB of files.<br><br>
<li>The <strong>troggle-generated reports</strong>: the tables of survey trips for each cave, the lists of scanned surveys for each survey trip, who was on what trip, calendar lists of logbook entries corrleated with recorded survey trips, who surveyed what caves, what source material was used to create the final cave survey images, who went on which expeditions (and who didn't go caving).
There are over 2,000 webpages generated using over 9,000 lines of bespoke python code from over 90 different urls.
The core of the troggle software is the data architecture: the set of tables into which all the cave survey and expo data is poured and stored. These tables are what enables us to produce a large number of different but consistent reports and views.
Also there have been tables added to the core representation which are not included in this old diagram, e.g. Scannedimage, Survexdirectory, Survexscansfolder, Survexscansingle, Tunnelfile, TunnelfileSurvexscansfolders, Survey. See <ahref="datamodel.html">Troggle data model</a> python code (15 May 2020) and click on the Class Diagram below on the right.
<p>The reasons why we have an online system at all are described in our <ahref="../website-history.html">website history</a>.
<p>There is an introductory article "<ahref="/expofiles/documents/troggle/troggle2020.pdf"download>Troggle: a revised system for cave data management</a>".
Troggle is written in Python (about 11,000 lines excluding comments) and is built on the Django framework. Before starting to work on Troggle it might be a good idea to run through an initial install and exploration of <ahref="https://code.djangoproject.com/wiki/Tutorials">a tutorial Django project</a> to get the Django concepts bedded down - which are not at all obvious and some exist only within Django.
<p>Django is the thing that puts the survey data in a database in a way that helps us write far less code to get it in and out again, and provides templates which make it quicker to maintain all the webpages. <ahref="http://www.djangoproject.com/"><imgclass="onright"src="https://www.djangoproject.com/m/img/badges/djangopowered126x54.gif"border="0"alt="Powered by Django."title="Powered by Django."/></a>
See the <ahref="https://docs.djangoproject.com/en/dev/misc/design-philosophies/">django design philosophy</a> for why we chose it: while django comes with a full stack (db, request/response, URL mapping, HTML templates) the layers of the stack are independent and individually replaceable.
<p>To understand how troggle imports the data from the survex files, tunnel files, logbooks etc., see the <ahref="trogimport.html">troggle import (databaseReset.py)</a> documentation.
<p>The following separate import operations are managed by the import utility <ahref="trogimport.html">(databaseReset.py)</a>:
<p>There are only four places where this happens. This is where online forms are used to create cave entrance records and cave records, where a form is used to record the information about a wallet. These are created in the database but also exported as files so that when troggle is rebuilt and data reimported the new cave data is there.
<p>Any page in this handbook, and any logbook entry, can also be edited online and the page is saved as a file - and registered with the version control system.