2
0
mirror of https://expo.survex.com/repositories/troggle/.git synced 2024-11-28 10:02:10 +00:00

Removing unneeded svx from parse list

This commit is contained in:
Philip Sargent 2023-09-08 01:26:01 +03:00
parent 5fc5c1285a
commit 28d1092956
6 changed files with 81 additions and 163 deletions

1
.gitignore vendored
View File

@ -72,3 +72,4 @@ pydebianbullseye
javascript javascript
_troggle_import_root.svx

View File

@ -10,31 +10,6 @@ from django.urls import reverse
# from troggle.core.models.troggle import DataIssue # circular import. Hmm # from troggle.core.models.troggle import DataIssue # circular import. Hmm
# class SurvexDirectory(models.Model):
# """This relates a survexfile (identified by path) to the primary SurvexFile
# which is the 'head' of the survex tree for that cave.
# Surely this could just be a property of Cave ? No. Several subdirectories
# all relate to the same Cave.
# But it *could* be a property of SurvexFile
# """
# path = models.CharField(max_length=200)
# primarysurvexfile = models.ForeignKey(
# "SurvexFile", related_name="primarysurvexfile", blank=True, null=True, on_delete=models.SET_NULL
# )
# class Meta:
# ordering = ("id",)
# verbose_name_plural = "Survex directories"
# def contents(self):
# return "[SvxDir:" + str(self.path) + " | Primary svx:" + str(self.primarysurvexfile.path) + ".svx ]"
# def __str__(self):
# return "[SvxDir:" + str(self.path)+ "]"
class SurvexFile(models.Model): class SurvexFile(models.Model):
path = models.CharField(max_length=200) path = models.CharField(max_length=200)
#survexdirectory = models.ForeignKey("SurvexDirectory", blank=True, null=True, on_delete=models.SET_NULL) #survexdirectory = models.ForeignKey("SurvexDirectory", blank=True, null=True, on_delete=models.SET_NULL)
@ -56,19 +31,6 @@ class SurvexFile(models.Model):
fname = Path(settings.SURVEX_DATA, self.path + ".svx") fname = Path(settings.SURVEX_DATA, self.path + ".svx")
return fname.is_file() return fname.is_file()
# def SetDirectory(self):
# dirpath = os.path.split(self.path)[0]
# # pointless search every time we import a survex file if we know there are no duplicates..
# # don't use this for initial import.
# survexdirectorylist = SurvexDirectory.objects.filter(cave=self.cave, path=dirpath)
# if survexdirectorylist:
# self.survexdirectory = survexdirectorylist[0]
# else:
# survexdirectory = SurvexDirectory(path=dirpath, cave=self.cave, primarysurvexfile=self)
# survexdirectory.save()
# self.survexdirectory = survexdirectory
# self.save()
# Don't change from the default as that breaks troggle webpages and internal referencing! # Don't change from the default as that breaks troggle webpages and internal referencing!
# def __str__(self): # def __str__(self):
# return "[SurvexFile:"+str(self.path) + "-" + str(self.survexdirectory) + "-" + str(self.cave)+"]" # return "[SurvexFile:"+str(self.path) + "-" + str(self.survexdirectory) + "-" + str(self.cave)+"]"
@ -97,17 +59,6 @@ class SurvexStation(models.Model):
y = models.FloatField(blank=True, null=True) y = models.FloatField(blank=True, null=True)
z = models.FloatField(blank=True, null=True) z = models.FloatField(blank=True, null=True)
# def path(self):
# r = self.name
# b = self.block
# while True:
# if b.name:
# r = b.name + "." + r
# if b.parent:
# b = b.parent
# else:
# return r
class Meta: class Meta:
ordering = ("id",) ordering = ("id",)
@ -119,7 +70,7 @@ class SurvexStation(models.Model):
import math import math
def utmToLatLng(zone, easting, northing, northernHemisphere=True): def utmToLatLng(zone, easting, northing, northernHemisphere=True): # move this to utils.py ?
if not northernHemisphere: if not northernHemisphere:
northing = 10000000 - northing northing = 10000000 - northing
@ -206,10 +157,10 @@ class SurvexBlock(models.Model):
date = models.DateField(blank=True, null=True) date = models.DateField(blank=True, null=True)
expedition = models.ForeignKey("Expedition", blank=True, null=True, on_delete=models.SET_NULL) expedition = models.ForeignKey("Expedition", blank=True, null=True, on_delete=models.SET_NULL)
# if the survexfile object is deleted, then all teh suvex-blocks in it should be too, # if the survexfile object is deleted, then all the suvex-blocks in it should be too,
# though a block can span more than one file... # though a block can span more than one file...
survexfile = models.ForeignKey("SurvexFile", blank=True, null=True, on_delete=models.CASCADE) survexfile = models.ForeignKey("SurvexFile", blank=True, null=True, on_delete=models.CASCADE)
survexpath = models.CharField(max_length=200) # the path for the survex stations # survexpath = models.CharField(max_length=200, blank=True, null=True) No need for this anymore
scanswallet = models.ForeignKey( scanswallet = models.ForeignKey(
"Wallet", null=True, on_delete=models.SET_NULL "Wallet", null=True, on_delete=models.SET_NULL
@ -221,9 +172,6 @@ class SurvexBlock(models.Model):
class Meta: class Meta:
ordering = ("id",) ordering = ("id",)
# def __str__(self):
# return "[SurvexBlock:" + str(self.name) + "-path:" + str(self.survexpath) + "-cave:" + str(self.cave) + "]"
def __str__(self): def __str__(self):
return self.name and str(self.name) or "no_name-#" + str(self.id) return self.name and str(self.name) or "no_name-#" + str(self.id)
@ -244,7 +192,6 @@ class SurvexBlock(models.Model):
if index not in range(0, mx): if index not in range(0, mx):
print(f"DayIndex: More than {mx-1} SurvexBlock items on one day '{index}' {self}, restarting colour sequence.") print(f"DayIndex: More than {mx-1} SurvexBlock items on one day '{index}' {self}, restarting colour sequence.")
index = index % mx index = index % mx
# return list(self.survexblock_set.all()).index(self)
return index return index
@ -253,7 +200,7 @@ class SurvexPersonRole(models.Model):
is deleted too is deleted too
""" """
survexblock = models.ForeignKey("SurvexBlock", on_delete=models.CASCADE) survexblock = models.ForeignKey("SurvexBlock", on_delete=models.CASCADE)
# increasing levels of precision, Surely we only need survexblock and person now that we have no link to a logbook entry? # increasing levels of precision, Surely we only need survexblock and (either person or personexpedition)?
personname = models.CharField(max_length=100) personname = models.CharField(max_length=100)
person = models.ForeignKey("Person", blank=True, null=True, on_delete=models.CASCADE) # not needed person = models.ForeignKey("Person", blank=True, null=True, on_delete=models.CASCADE) # not needed
personexpedition = models.ForeignKey("PersonExpedition", blank=True, null=True, on_delete=models.SET_NULL) personexpedition = models.ForeignKey("PersonExpedition", blank=True, null=True, on_delete=models.SET_NULL)

View File

@ -160,7 +160,7 @@ def GetTripPersons(trippeople, expedition, logtime_underground, tid=None):
guests.append(nickname_used) guests.append(nickname_used)
if known_foreigner(nickname_used): if known_foreigner(nickname_used):
message = f" ! - {expedition.year} Known foreigner: '{nickname_used}' in entry {tid=}" message = f" ! - {expedition.year} Known foreigner: '{nickname_used}' in entry {tid=}"
print(message) # print(message)
else: else:
message = f" ! - {expedition.year} No name match for: '{nickname_used}' in entry {tid=} for this year." message = f" ! - {expedition.year} No name match for: '{nickname_used}' in entry {tid=} for this year."
print(message) print(message)

View File

@ -4,7 +4,7 @@ import re
import subprocess import subprocess
import sys import sys
import time import time
from datetime import datetime, timezone from datetime import date, datetime, timezone
from pathlib import Path from pathlib import Path
import troggle.settings as settings import troggle.settings as settings
@ -44,9 +44,7 @@ todo = """
Also it will be a tiny bit more accurate as these leg lengths are after loop closure fixup. Also it will be a tiny bit more accurate as these leg lengths are after loop closure fixup.
""" """
survexblockroot = None survexblockroot = None
survexomitsroot = None
ROOTBLOCK = "rootblock" ROOTBLOCK = "rootblock"
OMITBLOCK = "omitblock"
METRESINFEET = 3.28084 METRESINFEET = 3.28084
UNSEENS = "_unseens.svx" UNSEENS = "_unseens.svx"
@ -211,7 +209,7 @@ def confirm_team_on_trip(survexblock):
def check_team_cache(): def check_team_cache():
global trip_team_cache global trip_team_cache
message = f"! *team CACHEFAIL check_team_cache() called " message = f"! check_team_cache() called.. "
print(message) print(message)
for block in trip_team_cache: for block in trip_team_cache:
@ -643,7 +641,7 @@ class LoadingSurvex:
expo = self.expos[year] expo = self.expos[year]
else: else:
expeditions = Expedition.objects.filter(year=year) expeditions = Expedition.objects.filter(year=year)
if len(expeditions) != 1: if len(expeditions) > 1:
message = ( message = (
f"! More than one expedition in year {year} '{line}' ({survexblock}) {survexblock.survexfile.path}" f"! More than one expedition in year {year} '{line}' ({survexblock}) {survexblock.survexfile.path}"
) )
@ -708,11 +706,13 @@ class LoadingSurvex:
perps = get_people_on_trip(survexblock) # What, you don't know Judge Dredd slang ? perps = get_people_on_trip(survexblock) # What, you don't know Judge Dredd slang ?
if len(line) > 10: if len(line) > 10:
message = "! DATE Warning LONG DATE '{}' ({}) {}".format(oline, survexblock, survexblock.survexfile.path)
print(self.insp+message)
stash_data_issue(parser='svxdate', message=message, url=None, sb=(survexblock.survexfile.path))
if line[10] == "-": # ie a range, just look at first date if line[10] == "-": # ie a range, just look at first date
line = line[0:10] line = line[0:10]
else:
message = f"! DATE Warning LONG DATE '{oline}' ({survexblock}) {survexblock.survexfile.path}"
print(self.insp+message)
stash_data_issue(parser='svxdate', message=message, url=None, sb=(survexblock.survexfile.path))
if len(line) == 10: if len(line) == 10:
year = line[:4] year = line[:4]
@ -1220,7 +1220,8 @@ class LoadingSurvex:
# It is too late to add it to the pending caves list here, they were already # It is too late to add it to the pending caves list here, they were already
# processed in parsers/caves.py So we have to do a bespoke creation. # processed in parsers/caves.py So we have to do a bespoke creation.
cave = create_new_cave(includelabel) svxpath= includelabel
cave = create_new_cave(svxpath)
message = f" ! Warning: cave identifier '{caveid}'or {id} (guessed from file path) is not a known cave. Need to add to expoweb/cave_data/pendingcaves.txt ? In '{includelabel}.svx' at depth:[{len(depth)}]." message = f" ! Warning: cave identifier '{caveid}'or {id} (guessed from file path) is not a known cave. Need to add to expoweb/cave_data/pendingcaves.txt ? In '{includelabel}.svx' at depth:[{len(depth)}]."
print("\n" + message) print("\n" + message)
@ -1228,10 +1229,6 @@ class LoadingSurvex:
print(f"{self.pending}", end="", file=sys.stderr) print(f"{self.pending}", end="", file=sys.stderr)
stash_data_issue(parser="survex", message=message, url=None, sb=(includelabel)) stash_data_issue(parser="survex", message=message, url=None, sb=(includelabel))
# It is too late to add it to pending caves here, they were already processed in parsers/caves.py
# and something else is creating them...
# cave = create_new_cave(includelabel)
def LoadSurvexFile(self, svxid): def LoadSurvexFile(self, svxid):
"""Creates SurvexFile in the database, and SurvexDirectory if needed """Creates SurvexFile in the database, and SurvexDirectory if needed
Creates a new current survexfile and valid .survexdirectory Creates a new current survexfile and valid .survexdirectory
@ -1408,7 +1405,7 @@ class LoadingSurvex:
expoyear = str(survexblock.date.year) expoyear = str(survexblock.date.year)
except: except:
print(f">> why is survexblock not set ?! in LoadSurvexQM()/n {survexblock.survexfile.path}") print(f">> why is survexblock not set ?! in LoadSurvexQM()/n {survexblock.survexfile.path}")
expoyear = "1970" expoyear = settings.EPOCH.year # 1970
@ -1667,7 +1664,7 @@ class LoadingSurvex:
newsurvexblock = SurvexBlock( newsurvexblock = SurvexBlock(
name=blkid, name=blkid,
parent=survexblock, parent=survexblock,
survexpath=pathlist, # survexpath=pathlist, # use the debug file, not this, for debugging
survexfile=self.currentsurvexfile, survexfile=self.currentsurvexfile,
legsall=0, legsall=0,
legslength=0.0, legslength=0.0,
@ -1913,10 +1910,6 @@ class LoadingSurvex:
if path in self.svxfileslist: if path in self.svxfileslist:
# We have already used os.normpath() so this is OK. "/../" and "//" have been simplified already. # We have already used os.normpath() so this is OK. "/../" and "//" have been simplified already.
if stop_dup_warning:
# print("D",end="", file=sys.stderr)
pass
else:
message = f" * Warning. Duplicate detected. We have already seen this *include '{path}' from another survex file. Detected at callcount:{self.callcount} depth:{self.depthinclude}" message = f" * Warning. Duplicate detected. We have already seen this *include '{path}' from another survex file. Detected at callcount:{self.callcount} depth:{self.depthinclude}"
print(message) print(message)
print(message, file=flinear) print(message, file=flinear)
@ -1926,7 +1919,7 @@ class LoadingSurvex:
message = f" ! ERROR. Should have been caught before this. Survex file already *included 2x. Probably an infinite loop so fix your *include statements that include this. Aborting. {path}" message = f" ! ERROR. Should have been caught before this. Survex file already *included 2x. Probably an infinite loop so fix your *include statements that include this. Aborting. {path}"
print(message) print(message)
print(message, file=flinear) print(message, file=flinear)
# print(message,file=sys.stderr) print(message,file=sys.stderr)
stash_data_issue(parser="survex", message=message, url=None, sb=(path)) stash_data_issue(parser="survex", message=message, url=None, sb=(path))
return return
return return
@ -2001,9 +1994,9 @@ class LoadingSurvex:
text=True, text=True,
) )
if sp.returncode != 0: if sp.returncode != 0:
message = f" ! Error running {settings.CAVERN}: {fullpath}" message = f" ! Error when running {settings.CAVERN}: {fullpath}"
url = f"/survexfile{fullpath}.svx".replace(str(settings.SURVEX_DATA), "") url = f"/survexfile{fullpath}.svx".replace(str(settings.SURVEX_DATA), "")
stash_data_issue(parser="xEntrances", message=message, url=url) stash_data_issue(parser="survex", message=message, url=url)
print(message) print(message)
print( print(
"stderr:\n\n" + str(sp.stderr) + "\n\n" + str(sp.stdout) + "\n\nreturn code: " + str(sp.returncode) "stderr:\n\n" + str(sp.stderr) + "\n\n" + str(sp.stdout) + "\n\nreturn code: " + str(sp.returncode)
@ -2069,21 +2062,37 @@ class LoadingSurvex:
runcavern() runcavern()
def FindAndLoadSurvex(survexblockroot): def FindAndLoadSurvex():
"""Follows the *include links successively to find survex files """Follows the *include links successively to find survex files
This proceeds in 3 phases: This proceeds in 3 phases:
1. The root survex file is read and all the *include files are found, using PushdownStackScan() 1. The root survex file is read and all the *include files are found, using PushdownStackScan()
2. All the other survex files in the :loser: repo are found, and their *includes found, 2. All the other survex files in the :loser: repo are found, and their *includes found,
using another PushdownStackScan() [duplicates omitted] using another PushdownStackScan() [duplicates omitted]
3. The combined expanded file containing all the survex data is parsed as a single file, 3. The combined expanded file containing all the survex data is parsed as a single file,
using LinearLoad()""" using LinearLoad()
global stop_dup_warning """
def make_survexblockroot():
survexfileroot = SurvexFile(
path=settings.SURVEX_TOPNAME, cave=None)
survexfileroot.save()
survexblockroot = SurvexBlock(
name=ROOTBLOCK, survexfile=survexfileroot, legsall=0, legslength=0.0)
# crashes here sometimes on MariaDB complaining that cave_id should not be null. But it should be.
# django.db.utils.IntegrityError: (1048, "Column 'cave_id' cannot be null")
# fix by restarting db on server
# sudo service mariadb stop
# sudo service mariadb start
survexblockroot.save()
return survexblockroot
print(" - redirecting stdout to svxblks.log...") print(" - redirecting stdout to svxblks.log...")
stdout_orig = sys.stdout stdout_orig = sys.stdout
# Redirect sys.stdout to the file # Redirect sys.stdout to the file
sys.stdout = open("svxblks.log", "w") sys.stdout = open("svxblks.log", "w")
print(f" - Scanning Survex Blocks tree from {settings.SURVEX_TOPNAME}.svx ...", file=sys.stderr) print(f" - Scanning Survex Blocks tree from {settings.SURVEX_TOPNAME}.svx ...", file=sys.stderr)
survexblockroot = make_survexblockroot()
survexfileroot = survexblockroot.survexfile # i.e. SURVEX_TOPNAME only survexfileroot = survexblockroot.survexfile # i.e. SURVEX_TOPNAME only
collatefilename = "_" + survexfileroot.path + ".svx" collatefilename = "_" + survexfileroot.path + ".svx"
@ -2169,24 +2178,29 @@ def FindAndLoadSurvex(survexblockroot):
file=sys.stderr, file=sys.stderr,
) )
# These exceptions WILL be parsed if the are *included by any file which is not excepted
unseensroot = re.sub(r"\.svx$", "", UNSEENS) unseensroot = re.sub(r"\.svx$", "", UNSEENS)
excpts = ["surface/terrain", "kataster/kataster-boundaries", "template", "docs", "deprecated", "subsections", unseensroot] excpts = ["surface/terrain", "kataster/kataster-boundaries", "gpx/gpx_publish/essentials", "template", "docs", "deprecated", "subsections", "1623-and-1626-no-schoenberg-hs", "1623-and-1624-and-1626-and-1627", "1623-and-1626",unseensroot]
removals = [] removals = set()
for x in unseens: for x in unseens:
for o in excpts: for o in excpts:
if str(x).strip().startswith(o): if str(x).strip().startswith(o):
removals.append(x) removals.add(x)
# special fix for .svx file not actually in survex format # special fix for .svx file not actually in survex format
unseens.remove(Path("fixedpts/gps/gps00raw")) unseens.remove(Path("fixedpts/gps/gps00raw"))
for x in removals: for x in removals:
unseens.remove(x) unseens.remove(x)
print(f" x NOT parsing {x}")
print( print(
f"\n - {len(unseens)} survex files found which were not included in main tree. ({len(svx_scan.svxfileslist)} in main tree)", f"\n - {len(unseens)} survex files found which were not included in main tree. ({len(svx_scan.svxfileslist)} in main tree)",
file=sys.stderr, file=sys.stderr,
) )
print(f" -- (but ignoring {len(removals)} of them)", file=sys.stderr)
check_team_cache() check_team_cache()
print(f" -- Now loading the previously-omitted survex files.", file=sys.stderr) s_date = date.today().isoformat().replace('-','.')
print(f" -- Now loading the previously-omitted survex files as {UNSEENS} *date {s_date}", file=sys.stderr)
print(f" - (except: {excpts})", file=sys.stderr) print(f" - (except: {excpts})", file=sys.stderr)
with open(Path(settings.SURVEX_DATA, UNSEENS), "w") as u: with open(Path(settings.SURVEX_DATA, UNSEENS), "w") as u:
@ -2196,6 +2210,8 @@ def FindAndLoadSurvex(survexblockroot):
u.write(f"; autogenerated by parser/survex.py from databasereset.py on '{datetime.now(timezone.utc)}'\n") u.write(f"; autogenerated by parser/survex.py from databasereset.py on '{datetime.now(timezone.utc)}'\n")
u.write(f"; omitting any file beginning with {excpts}\n\n") u.write(f"; omitting any file beginning with {excpts}\n\n")
u.write("*begin troggle_unseens\n") u.write("*begin troggle_unseens\n")
u.write("*team something Nobody\n")
u.write(f"*date {s_date}\n")
u.write("*title \"Collated unseen and unlinked survex files\"\n") u.write("*title \"Collated unseen and unlinked survex files\"\n")
for x in sorted(unseens): for x in sorted(unseens):
u.write(f" *include {x}\n") u.write(f" *include {x}\n")
@ -2227,11 +2243,9 @@ def FindAndLoadSurvex(survexblockroot):
fcollate.write(f";*include {UNSEENS}\n") fcollate.write(f";*include {UNSEENS}\n")
flinear.write(f"{omit_scan.depthinclude:2} {indent} *include {unseensroot}\n") flinear.write(f"{omit_scan.depthinclude:2} {indent} *include {unseensroot}\n")
omit_scan.svxpass = omit_scan.ODDS omit_scan.svxpass = omit_scan.ODDS
# stop_dup_warning = True
# ---------------------------------------------------------------- # ----------------------------------------------------------------
omit_scan.PushdownStackScan(survexblockroot, unseensroot, finrootname, flinear, fcollate) omit_scan.PushdownStackScan(survexblockroot, unseensroot, finrootname, flinear, fcollate)
# ---------------------------------------------------------------- # ----------------------------------------------------------------
# stop_dup_warning = False
omit_scan.svxpass = "" omit_scan.svxpass = ""
flinear.write(f"{omit_scan.depthinclude:2} {indent} *edulcni {unseensroot}\n") flinear.write(f"{omit_scan.depthinclude:2} {indent} *edulcni {unseensroot}\n")
@ -2269,10 +2283,6 @@ def FindAndLoadSurvex(survexblockroot):
print("\n - Loading All Survex Blocks (LinearLoad)", file=sys.stderr) print("\n - Loading All Survex Blocks (LinearLoad)", file=sys.stderr)
svx_load = LoadingSurvex() svx_load = LoadingSurvex()
#svx_load.survexdict[survexfileroot.survexdirectory] = []
#svx_load.survexdict[survexfileroot.survexdirectory].append(survexfileroot)
#svx_load.svxdirs[""] = survexfileroot.survexdirectory
# pr2 = cProfile.Profile() # pr2 = cProfile.Profile()
# pr2.enable() # pr2.enable()
print(" ", file=sys.stderr, end="") print(" ", file=sys.stderr, end="")
@ -2298,7 +2308,7 @@ def FindAndLoadSurvex(survexblockroot):
mem1 = get_process_memory() mem1 = get_process_memory()
print(f" - Number of SurvexDirectories: {len(svx_load.svxprim):,}") print(f" - Number of SurvexDirectories: {len(svx_load.svxprim):,}")
tf = SurvexFile.objects.all().count() tf = SurvexFile.objects.all().count() - len(removals)
print(f" - Number of SurvexFiles: {tf:,}") print(f" - Number of SurvexFiles: {tf:,}")
print(f" - Number of Survex legs: {legsnumber:,}") print(f" - Number of Survex legs: {legsnumber:,}")
svx_load = None svx_load = None
@ -2328,6 +2338,9 @@ def parse_one_file(fpath): # --------------------------------------in progress--
"""Parse just one file. Use when re-loading after editing. """Parse just one file. Use when re-loading after editing.
NOTE: *include lines are ignored. NOTE: *include lines are ignored.
But this is fine, they will already be in the system, UNLESS a new *include line is edited in
without also opening that file in the online editor.
In the initial file parsing in databaseReset, the *include expansion is done In the initial file parsing in databaseReset, the *include expansion is done
in an earlier stange than LinearLoad(). By the time LinearLoad() is called, in an earlier stange than LinearLoad(). By the time LinearLoad() is called,
all the *include expansion has happened. all the *include expansion has happened.
@ -2348,7 +2361,7 @@ def parse_one_file(fpath): # --------------------------------------in progress--
# It is vital that the block has attached the survexfile object which is being parsed. # It is vital that the block has attached the survexfile object which is being parsed.
block_dummy = SurvexBlock( block_dummy = SurvexBlock(
name="dummy", survexpath="", survexfile=svxfileroot, legsall=0, legslength=0.0 name="dummy", survexfile=svxfileroot, legsall=0, legslength=0.0
) )
svxfileroot.save() svxfileroot.save()
block_dummy.save() block_dummy.save()
@ -2357,11 +2370,6 @@ def parse_one_file(fpath): # --------------------------------------in progress--
block_dummy.save() block_dummy.save()
print(f" - block_dummy now '{block_dummy}' {type(block_dummy)} id={block_dummy.id} f:{block_dummy.survexfile}") print(f" - block_dummy now '{block_dummy}' {type(block_dummy)} id={block_dummy.id} f:{block_dummy.survexfile}")
# svx_load.survexdict[svxfileroot.survexdirectory] = []
# svx_load.survexdict[svxfileroot.survexdirectory].append(svxfileroot)
#svx_load.svxdirs[""] = svxfileroot.survexdirectory
# ---------------------------------------------------------------- # ----------------------------------------------------------------
svx_load.LinearLoad(block_dummy, svxfileroot.path, fname) svx_load.LinearLoad(block_dummy, svxfileroot.path, fname)
# ---------------------------------------------------------------- # ----------------------------------------------------------------
@ -2441,31 +2449,21 @@ def parse_one_file(fpath): # --------------------------------------in progress--
svx_load = None svx_load = None
return True return True
def MakeSurvexFileRoot(): def MakeFileRoot(svxpath):
"""Returns a file_object.path = SURVEX_TOPNAME associated with directory_object.path = SURVEX_DATA""" """Returns a file_object.path
# find a cave, any cave.. Used by the online survex file editor when re-parsing
smk = Cave.objects.filter(kataster_number="000") # returns a list, a QuerySet
fileroot = SurvexFile(path=settings.SURVEX_TOPNAME, cave=None)
fileroot.save()
return fileroot
def MakeFileRoot(fn):
"""Returns a file_object.path = _unseens.svx associated with directory_object.path = SURVEX_DATA
or tries to find the primary survex file for this cave or tries to find the primary survex file for this cave
""" """
cave = IdentifyCave(fn) cave = IdentifyCave(svxpath)
if not cave: if not cave:
if fn != UNSEENS: if svxpath != UNSEENS:
cave = create_new_cave(fn) cave = create_new_cave(svxpath)
print(f" - Making/finding a new root survexfile for this import: {fn}") # is this really necessayr ?!
fileroot = SurvexFile(path=svxpath, cave=cave)
fileroot = SurvexFile(path=fn, cave=cave)
fileroot.save() fileroot.save()
fileroot.cave = cave print(f" - Making/finding a new dummy root survexfile for this import: {svxpath} with cave {cave}")
print(f" - new fileroot {type(fileroot)} for {fn} with cave {cave} - {fileroot}") print(f" - new fileroot {type(fileroot)} for {svxpath} with cave {cave}\n - {fileroot.primary} {fileroot.path} {fileroot.cave} ")
return fileroot return fileroot
@ -2501,35 +2499,14 @@ def LoadSurvexBlocks():
print(" - survex Data Issues flushed") print(" - survex Data Issues flushed")
mem1 = get_process_memory() mem1 = get_process_memory()
print(f" - MEM:{mem1:7.2f} MB now ", file=sys.stderr) print(f" - MEM:{mem1:7.2f} MB now ", file=sys.stderr)
survexfileroot = MakeSurvexFileRoot()
# this next makes a block_object assciated with a file_object.path = SURVEX_TOPNAME
survexblockroot = SurvexBlock(
name=ROOTBLOCK, survexpath="", survexfile=survexfileroot, legsall=0, legslength=0.0
)
# crashes here sometimes on MariaDB complaining that cave_id should not be null. But it should be.
# django.db.utils.IntegrityError: (1048, "Column 'cave_id' cannot be null")
# fix by restarting db on server
# sudo service mariadb stop
# sudo service mariadb start
survexblockroot.save()
omitsfileroot = MakeFileRoot(UNSEENS)
survexomitsroot = SurvexBlock(
name=OMITBLOCK, survexpath="", survexfile=omitsfileroot, legsall=0, legslength=0.0
)
survexomitsroot.save()
print(" - Loading Survex Blocks...") print(" - Loading Survex Blocks...")
memstart = get_process_memory() memstart = get_process_memory()
# ---------------------------------------------------------------- # ----------------------------------------------------------------
FindAndLoadSurvex(survexblockroot) FindAndLoadSurvex()
# ---------------------------------------------------------------- # ----------------------------------------------------------------
memend = get_process_memory() memend = get_process_memory()
print(f" - MEMORY start:{memstart:.3f} MB end:{memend:.3f} MB increase={memend - memstart:.3f} MB") print(f" - MEMORY start:{memstart:.3f} MB end:{memend:.3f} MB increase={memend - memstart:.3f} MB")
survexblockroot.save()
global person_pending_cache global person_pending_cache
for sb in person_pending_cache: for sb in person_pending_cache:
if len(person_pending_cache[sb]) > 0: if len(person_pending_cache[sb]) > 0:

View File

@ -74,15 +74,8 @@ USE_L10N = True
FIX_PERMISSIONS = [] FIX_PERMISSIONS = []
# top-level survex file basename (without .svx) # top-level survex file basename (without .svx)
SURVEX_TOPNAME = "1623-and-1626-no-schoenberg-hs" # SURVEX_TOPNAME = "1623-and-1626-no-schoenberg-hs"
SURVEX_TOPNAME = "troggle_import_root" # same, but without all the 'essentials' gubbins
# Caves for which survex files exist, but are not otherwise registered
# replaced (?) by expoweb/cave_data/pendingcaves.txt
# PENDING = ["1626-361", "2007-06", "2009-02",
# "2012-ns-01", "2012-ns-02", "2010-04", "2012-ns-05", "2012-ns-06",
# "2012-ns-07", "2012-ns-08", "2012-ns-12", "2012-ns-14", "2012-ns-15", "2014-bl888",
# "2018-pf-01", "2018-pf-02"]
APPEND_SLASH = ( APPEND_SLASH = (
False # never relevant because we have urls that match unknown files and produce an 'edit this page' response False # never relevant because we have urls that match unknown files and produce an 'edit this page' response

View File

@ -84,7 +84,7 @@ underground survey length: {{svxlength|floatformat:2}} metres<br />
cave primary survexfile <a href="/survexfile/{{ survexfile.cave.survex_file }}">{{survexfile.cave.survex_file}}</a> <br> cave primary survexfile <a href="/survexfile/{{ survexfile.cave.survex_file }}">{{survexfile.cave.survex_file}}</a> <br>
directory primary survexfile <a href="/survexfile/{{survexfile.primary}}.svx">{{survexfile.primary}}</a> <br /> directory primary survexfile <a href="/survexfile/{{survexfile.primary}}.svx">{{survexfile.primary}}</a> <br />
{% for sb in svxblocks %} {% for sb in svxblocks %}
block <em>{{sb}}</em> has parent block <em>{{sb.parent}}</em><br /> block <em>{{sb}}</em> has parent block <em>{{sb.parent}}</em> (in file {{sb.parent.survexfile.path}}.svx)<br />
{% empty %} {% empty %}
Cannot find any <em>dated</em> survex blocks in this survex file (not looking at *include files). <br /> Cannot find any <em>dated</em> survex blocks in this survex file (not looking at *include files). <br />
Report this to a nerd if you think this is incorrect. Report this to a nerd if you think this is incorrect.