1 |
# Copyright (C) 2001, 2002 by Intevation GmbH |
# Copyright (C) 2001, 2002, 2003, 2004 by Intevation GmbH |
2 |
# Authors: |
# Authors: |
3 |
# Jan-Oliver Wagner <[email protected]> |
# Jan-Oliver Wagner <[email protected]> |
4 |
# Bernhard Herzog <[email protected]> |
# Bernhard Herzog <[email protected]> |
13 |
|
|
14 |
__version__ = "$Revision$" |
__version__ = "$Revision$" |
15 |
|
|
16 |
import sys, string, os |
import string, os |
17 |
|
|
18 |
import xml.sax |
import xml.sax |
19 |
import xml.sax.handler |
import xml.sax.handler |
20 |
from xml.sax import make_parser, ErrorHandler |
from xml.sax import make_parser, ErrorHandler, SAXNotRecognizedException |
21 |
|
|
22 |
from Thuban import _ |
from Thuban import _ |
23 |
from Thuban.common import * |
|
24 |
|
from Thuban.Model.table import FIELDTYPE_INT, FIELDTYPE_DOUBLE, \ |
25 |
|
FIELDTYPE_STRING |
26 |
|
|
27 |
|
from Thuban.Model.color import Color, Transparent |
28 |
|
|
29 |
from Thuban.Model.session import Session |
from Thuban.Model.session import Session |
30 |
from Thuban.Model.map import Map |
from Thuban.Model.map import Map |
31 |
from Thuban.Model.layer import Layer |
from Thuban.Model.layer import Layer, RasterLayer |
|
from Thuban.Model.color import Color |
|
32 |
from Thuban.Model.proj import Projection |
from Thuban.Model.proj import Projection |
33 |
|
from Thuban.Model.range import Range |
34 |
from Thuban.Model.classification import Classification, \ |
from Thuban.Model.classification import Classification, \ |
35 |
ClassGroupDefault, ClassGroupSingleton, ClassGroupRange, ClassGroupMap, \ |
ClassGroupDefault, ClassGroupSingleton, ClassGroupRange, ClassGroupMap, \ |
36 |
ClassGroupProperties |
ClassGroupProperties |
37 |
|
from Thuban.Model.data import DerivedShapeStore, ShapefileStore |
38 |
|
from Thuban.Model.table import DBFTable |
39 |
|
from Thuban.Model.transientdb import TransientJoinedTable |
40 |
|
|
41 |
|
from Thuban.Model.xmlreader import XMLReader |
42 |
|
import resource |
43 |
|
|
44 |
|
import postgisdb |
45 |
|
|
46 |
|
class LoadError(Exception): |
47 |
|
|
48 |
|
"""Exception raised when the thuban file is corrupted |
49 |
|
|
50 |
|
Not all cases of corrupted thuban files will lead to this exception |
51 |
|
but those that are found by checks in the loading code itself are. |
52 |
|
""" |
53 |
|
|
54 |
|
|
55 |
|
class LoadCancelled(Exception): |
56 |
|
|
57 |
|
"""Exception raised to indicate that loading was interrupted by the user""" |
58 |
|
|
59 |
|
|
60 |
def parse_color(color): |
def parse_color(color): |
65 |
""" |
""" |
66 |
color = string.strip(color) |
color = string.strip(color) |
67 |
if color == "None": |
if color == "None": |
68 |
result = Color.None |
result = Transparent |
69 |
elif color[0] == '#': |
elif color[0] == '#': |
70 |
if len(color) == 7: |
if len(color) == 7: |
71 |
r = string.atoi(color[1:3], 16) / 255.0 |
r = string.atoi(color[1:3], 16) / 255.0 |
79 |
raise ValueError(_("Invalid color specification %s") % color) |
raise ValueError(_("Invalid color specification %s") % color) |
80 |
return result |
return result |
81 |
|
|
82 |
|
class AttrDesc: |
83 |
|
|
84 |
|
def __init__(self, name, required = False, default = "", |
85 |
|
conversion = None): |
86 |
|
if not isinstance(name, tuple): |
87 |
|
fullname = (None, name) |
88 |
|
else: |
89 |
|
fullname = name |
90 |
|
name = name[1] |
91 |
|
self.name = name |
92 |
|
self.fullname = fullname |
93 |
|
self.required = required |
94 |
|
self.default = default |
95 |
|
self.conversion = conversion |
96 |
|
|
97 |
class ProcessSession(xml.sax.handler.ContentHandler): |
# set by the SessionLoader's check_attrs method |
98 |
|
self.value = None |
99 |
|
|
100 |
|
|
101 |
|
class SessionLoader(XMLReader): |
102 |
|
|
103 |
|
def __init__(self, db_connection_callback = None): |
104 |
|
"""Inititialize the Sax handler.""" |
105 |
|
XMLReader.__init__(self) |
106 |
|
|
107 |
|
self.db_connection_callback = db_connection_callback |
108 |
|
|
|
# Dictionary mapping element names (or (URI, element name) pairs for |
|
|
# documents using namespaces) to method names. The methods should |
|
|
# accept the same parameters as the startElement (or startElementNS) |
|
|
# methods. The start_dispatcher is used by the default startElement |
|
|
# and startElementNS methods to call a method for the open tag of an |
|
|
# element. |
|
|
start_dispatcher = {} |
|
|
|
|
|
# end_dispatcher works just like start_dispatcher but it's used by |
|
|
# endElement and endElementNS. The method whose names it maps to |
|
|
# should accept the same parameters as endElement and endElementNS. |
|
|
end_dispatcher = {} |
|
|
|
|
|
|
|
|
def __init__(self, directory): |
|
|
"""Inititialize the Sax handler. |
|
|
|
|
|
The directory parameter should be the directory containing the |
|
|
session file. It's needed to interpret embedded relative |
|
|
filenames. |
|
|
""" |
|
|
self.directory = directory |
|
|
self.chars = '' |
|
109 |
self.theSession = None |
self.theSession = None |
110 |
self.aMap = None |
self.aMap = None |
111 |
self.aLayer = None |
self.aLayer = None |
112 |
|
|
113 |
def startElementNS(self, name, qname, attrs): |
# Map ids used in the thuban file to the corresponding objects |
114 |
"""Call the method given for name in self.start_dispatcher |
# in the session |
115 |
""" |
self.idmap = {} |
116 |
if name[0] is None: |
|
117 |
method_name = self.start_dispatcher.get(name[1]) |
dispatchers = { |
118 |
else: |
'session' : ("start_session", "end_session"), |
119 |
# Dispatch with namespace |
|
120 |
method_name = self.start_dispatcher.get(name) |
'dbconnection': ("start_dbconnection", None), |
121 |
if method_name is not None: |
|
122 |
getattr(self, method_name)(name, qname, attrs) |
'dbshapesource': ("start_dbshapesource", None), |
123 |
|
'fileshapesource': ("start_fileshapesource", None), |
124 |
def endElementNS(self, name, qname): |
'derivedshapesource': ("start_derivedshapesource", None), |
125 |
"""Call the method given for name in self.end_dispatcher |
'filetable': ("start_filetable", None), |
126 |
|
'jointable': ("start_jointable", None), |
127 |
|
|
128 |
|
'map' : ("start_map", "end_map"), |
129 |
|
'projection' : ("start_projection", "end_projection"), |
130 |
|
'parameter' : ("start_parameter", None), |
131 |
|
'layer' : ("start_layer", "end_layer"), |
132 |
|
'rasterlayer' : ("start_rasterlayer", "end_rasterlayer"), |
133 |
|
'classification': ("start_classification", "end_classification"), |
134 |
|
'clnull' : ("start_clnull", "end_clnull"), |
135 |
|
'clpoint' : ("start_clpoint", "end_clpoint"), |
136 |
|
'clrange' : ("start_clrange", "end_clrange"), |
137 |
|
'cldata' : ("start_cldata", "end_cldata"), |
138 |
|
'table' : ("start_table", "end_table"), |
139 |
|
'labellayer' : ("start_labellayer", None), |
140 |
|
'label' : ("start_label", None)} |
141 |
|
|
142 |
|
# all dispatchers should be used for the 0.8 and 0.9 namespaces too |
143 |
|
for xmlns in ("http://thuban.intevation.org/dtds/thuban-0.8.dtd", |
144 |
|
"http://thuban.intevation.org/dtds/thuban-0.9-dev.dtd", |
145 |
|
"http://thuban.intevation.org/dtds/thuban-0.9.dtd", |
146 |
|
"http://thuban.intevation.org/dtds/thuban-1.0-dev.dtd", |
147 |
|
"http://thuban.intevation.org/dtds/thuban-1.0rc1.dtd", |
148 |
|
"http://thuban.intevation.org/dtds/thuban-1.0.0.dtd", |
149 |
|
"http://thuban.intevation.org/dtds/thuban-1.1-dev.dtd"): |
150 |
|
for key, value in dispatchers.items(): |
151 |
|
dispatchers[(xmlns, key)] = value |
152 |
|
|
153 |
|
XMLReader.AddDispatchers(self, dispatchers) |
154 |
|
|
155 |
|
def Destroy(self): |
156 |
|
"""Clear all instance variables to cut cyclic references. |
157 |
|
|
158 |
|
The GC would have collected the loader eventually but it can |
159 |
|
happen that it doesn't run at all until Thuban is closed (2.3 |
160 |
|
but not 2.2 tries a bit harder and forces a collection when the |
161 |
|
interpreter terminates) |
162 |
""" |
""" |
163 |
if name[0] is None: |
self.__dict__.clear() |
|
method_name = self.end_dispatcher.get(name[1]) |
|
|
else: |
|
|
# Dispatch with namespace |
|
|
method_name = self.end_dispatcher.get(name) |
|
|
if method_name is not None: |
|
|
getattr(self, method_name)(name, qname) |
|
164 |
|
|
165 |
def start_session(self, name, qname, attrs): |
def start_session(self, name, qname, attrs): |
166 |
self.theSession = Session(attrs.get((None, 'title'), None)) |
self.theSession = Session(self.encode(attrs.get((None, 'title'), |
167 |
start_dispatcher['session'] = "start_session" |
None))) |
168 |
|
|
169 |
def end_session(self, name, qname): |
def end_session(self, name, qname): |
170 |
pass |
pass |
171 |
end_dispatcher['session'] = "end_session" |
|
172 |
|
def check_attrs(self, element, attrs, descr): |
173 |
|
"""Check and convert some of the attributes of an element |
174 |
|
|
175 |
|
Parameters: |
176 |
|
element -- The element name |
177 |
|
attrs -- The attrs mapping as passed to the start_* methods |
178 |
|
descr -- Sequence of attribute descriptions (AttrDesc instances) |
179 |
|
|
180 |
|
Return a dictionary containig normalized versions of the |
181 |
|
attributes described in descr. The keys of that dictionary are |
182 |
|
the name attributes of the attribute descriptions. The attrs |
183 |
|
dictionary will not be modified. |
184 |
|
|
185 |
|
If the attribute is required, i.e. the 'required' attribute of |
186 |
|
the descrtiption is true, but it is not in attrs, raise a |
187 |
|
LoadError. |
188 |
|
|
189 |
|
If the attribute has a default value and it is not present in |
190 |
|
attrs, use that default value as the value in the returned dict. |
191 |
|
|
192 |
|
The value is converted before putting it into the returned dict. |
193 |
|
The following conversions are available: |
194 |
|
|
195 |
|
'filename' -- The attribute is a filename. |
196 |
|
|
197 |
|
If the filename is a relative name, interpret |
198 |
|
it relative to the directory containing the |
199 |
|
.thuban file and make it an absolute name |
200 |
|
|
201 |
|
'shapestore' -- The attribute is the ID of a shapestore |
202 |
|
defined earlier in the .thuban file. Look it |
203 |
|
up self.idmap |
204 |
|
|
205 |
|
'table' -- The attribute is the ID of a table or shapestore |
206 |
|
defined earlier in the .thuban file. Look it up |
207 |
|
self.idmap. If it's the ID of a shapestore the |
208 |
|
value will be the table of the shapestore. |
209 |
|
|
210 |
|
'idref' -- The attribute is the id of an object defined |
211 |
|
earlier in the .thuban file. Look it up self.idmap |
212 |
|
|
213 |
|
'ascii' -- The attribute is converted to a bytestring with |
214 |
|
ascii encoding. |
215 |
|
|
216 |
|
a callable -- The attribute value is passed to the callable |
217 |
|
and the return value is used as the converted |
218 |
|
value |
219 |
|
|
220 |
|
If no conversion is specified for an attribute it is converted |
221 |
|
with self.encode. |
222 |
|
""" |
223 |
|
normalized = {} |
224 |
|
|
225 |
|
for d in descr: |
226 |
|
if d.required and not attrs.has_key(d.fullname): |
227 |
|
raise LoadError("Element %s requires an attribute %r" |
228 |
|
% (element, d.name)) |
229 |
|
value = attrs.get(d.fullname, d.default) |
230 |
|
|
231 |
|
if d.conversion in ("idref", "shapesource"): |
232 |
|
if value in self.idmap: |
233 |
|
value = self.idmap[value] |
234 |
|
else: |
235 |
|
raise LoadError("Element %s requires an already defined ID" |
236 |
|
" in attribute %r" |
237 |
|
% (element, d.name)) |
238 |
|
elif d.conversion == "table": |
239 |
|
if value in self.idmap: |
240 |
|
value = self.idmap[value] |
241 |
|
if isinstance(value, ShapefileStore): |
242 |
|
value = value.Table() |
243 |
|
else: |
244 |
|
raise LoadError("Element %s requires an already defined ID" |
245 |
|
" in attribute %r" |
246 |
|
% (element, d.name)) |
247 |
|
elif d.conversion == "filename": |
248 |
|
value = os.path.abspath(os.path.join(self.GetDirectory(), |
249 |
|
value)) |
250 |
|
elif d.conversion == "ascii": |
251 |
|
value = value.encode("ascii") |
252 |
|
elif d.conversion: |
253 |
|
# Assume it's a callable |
254 |
|
value = d.conversion(value) |
255 |
|
else: |
256 |
|
value = self.encode(value) |
257 |
|
|
258 |
|
normalized[d.name] = value |
259 |
|
return normalized |
260 |
|
|
261 |
|
def start_dbconnection(self, name, qname, attrs): |
262 |
|
attrs = self.check_attrs(name, attrs, |
263 |
|
[AttrDesc("id", True), |
264 |
|
AttrDesc("dbtype", True), |
265 |
|
AttrDesc("host", False, ""), |
266 |
|
AttrDesc("port", False, ""), |
267 |
|
AttrDesc("user", False, ""), |
268 |
|
AttrDesc("dbname", True)]) |
269 |
|
ID = attrs["id"] |
270 |
|
dbtype = attrs["dbtype"] |
271 |
|
if dbtype != "postgis": |
272 |
|
raise LoadError("dbtype %r not supported" % filetype) |
273 |
|
|
274 |
|
del attrs["id"] |
275 |
|
del attrs["dbtype"] |
276 |
|
|
277 |
|
# Try to open the connection and if it fails ask the user for |
278 |
|
# the correct parameters repeatedly. |
279 |
|
# FIXME: it would be better not to insist on getting a |
280 |
|
# connection here. We should handle this more like the raster |
281 |
|
# images where the layers etc still are created but are not |
282 |
|
# drawn in case Thuban can't use the data for various reasons |
283 |
|
while 1: |
284 |
|
try: |
285 |
|
conn = postgisdb.PostGISConnection(**attrs) |
286 |
|
break |
287 |
|
except postgisdb.ConnectionError, val: |
288 |
|
if self.db_connection_callback is not None: |
289 |
|
attrs = self.db_connection_callback(attrs, str(val)) |
290 |
|
if attrs is None: |
291 |
|
raise LoadCancelled |
292 |
|
else: |
293 |
|
raise |
294 |
|
|
295 |
|
self.idmap[ID] = conn |
296 |
|
self.theSession.AddDBConnection(conn) |
297 |
|
|
298 |
|
def start_dbshapesource(self, name, qname, attrs): |
299 |
|
attrs = self.check_attrs(name, attrs, |
300 |
|
[AttrDesc("id", True), |
301 |
|
AttrDesc("dbconn", True, |
302 |
|
conversion = "idref"), |
303 |
|
AttrDesc("tablename", True, |
304 |
|
conversion = "ascii"), |
305 |
|
# id_column and geometry_column were |
306 |
|
# newly introduced with thuban-1.1.dtd |
307 |
|
# where they're required. Since we |
308 |
|
# support the older formats too we |
309 |
|
# have them optional here. |
310 |
|
AttrDesc("id_column", False, "gid", |
311 |
|
conversion = "ascii"), |
312 |
|
AttrDesc("geometry_column", False, |
313 |
|
conversion = "ascii")]) |
314 |
|
# The default value of geometry_column to use when instantiating |
315 |
|
# the db shapestore is None which we currently can't easily use |
316 |
|
# in check_attrs |
317 |
|
geometry_column = attrs["geometry_column"] |
318 |
|
if not geometry_column: |
319 |
|
geometry_column = None |
320 |
|
dbopen = self.theSession.OpenDBShapeStore |
321 |
|
self.idmap[attrs["id"]] = dbopen(attrs["dbconn"], attrs["tablename"], |
322 |
|
id_column = attrs["id_column"], |
323 |
|
geometry_column=geometry_column) |
324 |
|
|
325 |
|
def start_fileshapesource(self, name, qname, attrs): |
326 |
|
attrs = self.check_attrs(name, attrs, |
327 |
|
[AttrDesc("id", True), |
328 |
|
AttrDesc("filename", True, |
329 |
|
conversion = "filename"), |
330 |
|
AttrDesc("filetype", True)]) |
331 |
|
ID = attrs["id"] |
332 |
|
filename = attrs["filename"] |
333 |
|
filetype = attrs["filetype"] |
334 |
|
if filetype != "shapefile": |
335 |
|
raise LoadError("shapesource filetype %r not supported" % filetype) |
336 |
|
self.idmap[ID] = self.theSession.OpenShapefile(filename) |
337 |
|
|
338 |
|
def start_derivedshapesource(self, name, qname, attrs): |
339 |
|
attrs = self.check_attrs(name, attrs, |
340 |
|
[AttrDesc("id", True), |
341 |
|
AttrDesc("shapesource", True, |
342 |
|
conversion = "shapesource"), |
343 |
|
AttrDesc("table", True, conversion="table")]) |
344 |
|
store = DerivedShapeStore(attrs["shapesource"], attrs["table"]) |
345 |
|
self.theSession.AddShapeStore(store) |
346 |
|
self.idmap[attrs["id"]] = store |
347 |
|
|
348 |
|
def start_filetable(self, name, qname, attrs): |
349 |
|
attrs = self.check_attrs(name, attrs, |
350 |
|
[AttrDesc("id", True), |
351 |
|
AttrDesc("title", True), |
352 |
|
AttrDesc("filename", True, |
353 |
|
conversion = "filename"), |
354 |
|
AttrDesc("filetype")]) |
355 |
|
filetype = attrs["filetype"] |
356 |
|
if filetype != "DBF": |
357 |
|
raise LoadError("shapesource filetype %r not supported" % filetype) |
358 |
|
table = DBFTable(attrs["filename"]) |
359 |
|
table.SetTitle(attrs["title"]) |
360 |
|
self.idmap[attrs["id"]] = self.theSession.AddTable(table) |
361 |
|
|
362 |
|
def start_jointable(self, name, qname, attrs): |
363 |
|
attrs = self.check_attrs(name, attrs, |
364 |
|
[AttrDesc("id", True), |
365 |
|
AttrDesc("title", True), |
366 |
|
AttrDesc("left", True, conversion="table"), |
367 |
|
AttrDesc("leftcolumn", True), |
368 |
|
AttrDesc("right", True, conversion="table"), |
369 |
|
AttrDesc("rightcolumn", True), |
370 |
|
|
371 |
|
# jointype is required for file |
372 |
|
# version 0.9 but this attribute |
373 |
|
# wasn't in the 0.8 version because of |
374 |
|
# an oversight so we assume it's |
375 |
|
# optional since we want to handle |
376 |
|
# both file format versions here. |
377 |
|
AttrDesc("jointype", False, |
378 |
|
default="INNER")]) |
379 |
|
|
380 |
|
jointype = attrs["jointype"] |
381 |
|
if jointype == "LEFT OUTER": |
382 |
|
outer_join = True |
383 |
|
elif jointype == "INNER": |
384 |
|
outer_join = False |
385 |
|
else: |
386 |
|
raise LoadError("jointype %r not supported" % jointype ) |
387 |
|
table = TransientJoinedTable(self.theSession.TransientDB(), |
388 |
|
attrs["left"], attrs["leftcolumn"], |
389 |
|
attrs["right"], attrs["rightcolumn"], |
390 |
|
outer_join = outer_join) |
391 |
|
table.SetTitle(attrs["title"]) |
392 |
|
self.idmap[attrs["id"]] = self.theSession.AddTable(table) |
393 |
|
|
394 |
def start_map(self, name, qname, attrs): |
def start_map(self, name, qname, attrs): |
395 |
"""Start a map.""" |
"""Start a map.""" |
396 |
self.aMap = Map(attrs.get((None, 'title'), None)) |
self.aMap = Map(self.encode(attrs.get((None, 'title'), None))) |
|
start_dispatcher['map'] = "start_map" |
|
397 |
|
|
398 |
def end_map(self, name, qname): |
def end_map(self, name, qname): |
399 |
self.theSession.AddMap(self.aMap) |
self.theSession.AddMap(self.aMap) |
400 |
end_dispatcher['map'] = "end_map" |
self.aMap = None |
401 |
|
|
402 |
def start_projection(self, name, qname, attrs): |
def start_projection(self, name, qname, attrs): |
403 |
self.ProjectionParams = [ ] |
attrs = self.check_attrs(name, attrs, |
404 |
start_dispatcher['projection'] = "start_projection" |
[AttrDesc("name", conversion=self.encode), |
405 |
|
AttrDesc("epsg", default=None, |
406 |
|
conversion=self.encode)]) |
407 |
|
self.projection_name = attrs["name"] |
408 |
|
self.projection_epsg = attrs["epsg"] |
409 |
|
self.projection_params = [ ] |
410 |
|
|
411 |
def end_projection(self, name, qname): |
def end_projection(self, name, qname): |
412 |
self.aMap.SetProjection(Projection(self.ProjectionParams)) |
if self.aLayer is not None: |
413 |
end_dispatcher['projection'] = "end_projection" |
obj = self.aLayer |
414 |
|
elif self.aMap is not None: |
415 |
|
obj = self.aMap |
416 |
|
else: |
417 |
|
assert False, "projection tag out of context" |
418 |
|
pass |
419 |
|
|
420 |
|
obj.SetProjection(Projection(self.projection_params, |
421 |
|
self.projection_name, |
422 |
|
epsg = self.projection_epsg)) |
423 |
|
|
424 |
def start_parameter(self, name, qname, attrs): |
def start_parameter(self, name, qname, attrs): |
425 |
s = attrs.get((None, 'value')) |
s = attrs.get((None, 'value')) |
426 |
s = str(s) # we can't handle unicode in proj |
s = str(s) # we can't handle unicode in proj |
427 |
self.ProjectionParams.append(s) |
self.projection_params.append(s) |
|
start_dispatcher['parameter'] = "start_parameter" |
|
428 |
|
|
429 |
def start_layer(self, name, qname, attrs, layer_class = Layer): |
def start_layer(self, name, qname, attrs, layer_class = Layer): |
430 |
"""Start a layer |
"""Start a layer |
433 |
attrs which may be a dictionary as well as the normal SAX attrs |
attrs which may be a dictionary as well as the normal SAX attrs |
434 |
object and bind it to self.aLayer. |
object and bind it to self.aLayer. |
435 |
""" |
""" |
436 |
title = attrs.get((None, 'title'), "") |
title = self.encode(attrs.get((None, 'title'), "")) |
437 |
filename = attrs.get((None, 'filename'), "") |
filename = attrs.get((None, 'filename'), "") |
438 |
filename = os.path.join(self.directory, filename) |
filename = os.path.join(self.GetDirectory(), filename) |
439 |
|
filename = self.encode(filename) |
440 |
|
visible = self.encode(attrs.get((None, 'visible'), "true")) != "false" |
441 |
fill = parse_color(attrs.get((None, 'fill'), "None")) |
fill = parse_color(attrs.get((None, 'fill'), "None")) |
442 |
stroke = parse_color(attrs.get((None, 'stroke'), "#000000")) |
stroke = parse_color(attrs.get((None, 'stroke'), "#000000")) |
443 |
stroke_width = int(attrs.get((None, 'stroke_width'), "1")) |
stroke_width = int(attrs.get((None, 'stroke_width'), "1")) |
444 |
self.aLayer = layer_class(title, filename, fill = fill, |
if attrs.has_key((None, "shapestore")): |
445 |
stroke = stroke, stroke_width = stroke_width) |
store = self.idmap[attrs[(None, "shapestore")]] |
446 |
start_dispatcher['layer'] = "start_layer" |
else: |
447 |
|
store = self.theSession.OpenShapefile(filename) |
448 |
|
self.aLayer = layer_class(title, store, |
449 |
|
fill = fill, stroke = stroke, |
450 |
|
lineWidth = stroke_width, |
451 |
|
visible = visible) |
452 |
|
|
453 |
def end_layer(self, name, qname): |
def end_layer(self, name, qname): |
454 |
self.aMap.AddLayer(self.aLayer) |
self.aMap.AddLayer(self.aLayer) |
455 |
end_dispatcher['layer'] = "end_layer" |
self.aLayer = None |
456 |
|
|
457 |
|
def start_rasterlayer(self, name, qname, attrs, layer_class = RasterLayer): |
458 |
|
title = self.encode(attrs.get((None, 'title'), "")) |
459 |
|
filename = attrs.get((None, 'filename'), "") |
460 |
|
filename = os.path.join(self.GetDirectory(), filename) |
461 |
|
filename = self.encode(filename) |
462 |
|
visible = self.encode(attrs.get((None, 'visible'), "true")) != "false" |
463 |
|
|
464 |
|
self.aLayer = layer_class(title, filename, visible = visible) |
465 |
|
|
466 |
|
def end_rasterlayer(self, name, qname): |
467 |
|
self.aMap.AddLayer(self.aLayer) |
468 |
|
self.aLayer = None |
469 |
|
|
470 |
def start_classification(self, name, qname, attrs): |
def start_classification(self, name, qname, attrs): |
471 |
self.aLayer.GetClassification().SetField( |
attrs = self.check_attrs(name, attrs, |
472 |
attrs.get((None, 'field'), None)) |
[AttrDesc("field", True), |
473 |
start_dispatcher['classification'] = "start_classification" |
AttrDesc("field_type", True)]) |
474 |
|
field = attrs["field"] |
475 |
|
fieldType = attrs["field_type"] |
476 |
|
|
477 |
|
dbFieldType = self.aLayer.GetFieldType(field) |
478 |
|
|
479 |
|
if fieldType != dbFieldType: |
480 |
|
raise ValueError(_("xml field type differs from database!")) |
481 |
|
|
482 |
|
# setup conversion routines depending on the kind of data |
483 |
|
# we will be seeing later on |
484 |
|
if fieldType == FIELDTYPE_STRING: |
485 |
|
self.conv = str |
486 |
|
elif fieldType == FIELDTYPE_INT: |
487 |
|
self.conv = lambda p: int(float(p)) |
488 |
|
elif fieldType == FIELDTYPE_DOUBLE: |
489 |
|
self.conv = float |
490 |
|
|
491 |
|
self.aLayer.SetClassificationColumn(field) |
492 |
|
|
493 |
def end_classification(self, name, qname): |
def end_classification(self, name, qname): |
494 |
pass |
pass |
|
end_dispatcher['classification'] = "end_classification" |
|
495 |
|
|
496 |
def start_clnull(self, name, qname, attrs): |
def start_clnull(self, name, qname, attrs): |
497 |
self.cl_group = ClassGroupDefault() |
self.cl_group = ClassGroupDefault() |
498 |
self.cl_group.SetLabel(attrs.get((None, 'label'), "")) |
self.cl_group.SetLabel(self.encode(attrs.get((None, 'label'), ""))) |
499 |
self.cl_prop = ClassGroupProperties() |
self.cl_prop = ClassGroupProperties() |
|
start_dispatcher['clnull'] = "start_clnull" |
|
500 |
|
|
501 |
def end_clnull(self, name, qname): |
def end_clnull(self, name, qname): |
502 |
self.cl_group.SetProperties(self.cl_prop) |
self.cl_group.SetProperties(self.cl_prop) |
503 |
self.aLayer.GetClassification().SetDefaultGroup(self.cl_group) |
self.aLayer.GetClassification().SetDefaultGroup(self.cl_group) |
504 |
del self.cl_group, self.cl_prop |
del self.cl_group, self.cl_prop |
|
end_dispatcher['clnull'] = "end_clnull" |
|
505 |
|
|
506 |
def start_clpoint(self, name, qname, attrs): |
def start_clpoint(self, name, qname, attrs): |
507 |
attrib_value = attrs.get((None, 'value'), "0") |
attrib_value = attrs.get((None, 'value'), "0") |
508 |
|
|
509 |
try: |
field = self.aLayer.GetClassificationColumn() |
510 |
value = Str2Num(attrib_value) |
if self.aLayer.GetFieldType(field) == FIELDTYPE_STRING: |
511 |
except: |
value = self.encode(attrib_value) |
512 |
value = attrib_value |
else: |
513 |
|
value = self.conv(attrib_value) |
514 |
self.cl_group = ClassGroupSingleton(value) |
self.cl_group = ClassGroupSingleton(value) |
515 |
self.cl_group.SetLabel(attrs.get((None, 'label'), "")) |
self.cl_group.SetLabel(self.encode(attrs.get((None, 'label'), ""))) |
516 |
self.cl_prop = ClassGroupProperties() |
self.cl_prop = ClassGroupProperties() |
517 |
|
|
|
start_dispatcher['clpoint'] = "start_clpoint" |
|
518 |
|
|
519 |
def end_clpoint(self, name, qname): |
def end_clpoint(self, name, qname): |
520 |
self.cl_group.SetProperties(self.cl_prop) |
self.cl_group.SetProperties(self.cl_prop) |
521 |
self.aLayer.GetClassification().AddGroup(self.cl_group) |
self.aLayer.GetClassification().AppendGroup(self.cl_group) |
522 |
del self.cl_group, self.cl_prop |
del self.cl_group, self.cl_prop |
|
end_dispatcher['clpoint'] = "end_clpoint" |
|
523 |
|
|
524 |
def start_clrange(self, name, qname, attrs): |
def start_clrange(self, name, qname, attrs): |
525 |
|
|
526 |
|
range = attrs.get((None, 'range'), None) |
527 |
|
# for backward compatibility (min/max are not saved) |
528 |
|
min = attrs.get((None, 'min'), None) |
529 |
|
max = attrs.get((None, 'max'), None) |
530 |
|
|
531 |
try: |
try: |
532 |
min = Str2Num(attrs.get((None, 'min'), "0")) |
if range is not None: |
533 |
max = Str2Num(attrs.get((None, 'max'), "0")) |
self.cl_group = ClassGroupRange(Range(range)) |
534 |
|
elif min is not None and max is not None: |
535 |
|
self.cl_group = ClassGroupRange((self.conv(min), |
536 |
|
self.conv(max))) |
537 |
|
else: |
538 |
|
self.cl_group = ClassGroupRange(Range(None)) |
539 |
|
|
540 |
except ValueError: |
except ValueError: |
541 |
raise ValueError(_("Classification range is not a number!")) |
raise ValueError(_("Classification range is not a number!")) |
542 |
|
|
|
self.cl_group = ClassGroupRange(min, max) |
|
543 |
self.cl_group.SetLabel(attrs.get((None, 'label'), "")) |
self.cl_group.SetLabel(attrs.get((None, 'label'), "")) |
544 |
self.cl_prop = ClassGroupProperties() |
self.cl_prop = ClassGroupProperties() |
545 |
|
|
|
start_dispatcher['clrange'] = "start_clrange" |
|
546 |
|
|
547 |
def end_clrange(self, name, qname): |
def end_clrange(self, name, qname): |
548 |
self.cl_group.SetProperties(self.cl_prop) |
self.cl_group.SetProperties(self.cl_prop) |
549 |
self.aLayer.GetClassification().AddGroup(self.cl_group) |
self.aLayer.GetClassification().AppendGroup(self.cl_group) |
550 |
del self.cl_group, self.cl_prop |
del self.cl_group, self.cl_prop |
|
end_dispatcher['clrange'] = "end_clrange" |
|
551 |
|
|
552 |
def start_cldata(self, name, qname, attrs): |
def start_cldata(self, name, qname, attrs): |
553 |
self.cl_prop.SetStroke(parse_color(attrs.get((None, 'stroke'), "None"))) |
self.cl_prop.SetLineColor( |
554 |
self.cl_prop.SetStrokeWidth( |
parse_color(attrs.get((None, 'stroke'), "None"))) |
555 |
|
self.cl_prop.SetLineWidth( |
556 |
int(attrs.get((None, 'stroke_width'), "0"))) |
int(attrs.get((None, 'stroke_width'), "0"))) |
557 |
|
self.cl_prop.SetSize(int(attrs.get((None, 'size'), "5"))) |
558 |
self.cl_prop.SetFill(parse_color(attrs.get((None, 'fill'), "None"))) |
self.cl_prop.SetFill(parse_color(attrs.get((None, 'fill'), "None"))) |
|
start_dispatcher['cldata'] = "start_cldata" |
|
559 |
|
|
560 |
def end_cldata(self, name, qname): |
def end_cldata(self, name, qname): |
561 |
pass |
pass |
|
end_dispatcher['cldata'] = "end_cldata" |
|
|
|
|
|
def start_table(self, name, qname, attrs): |
|
|
print "table title: %s" % attrs.get('title', None) |
|
|
start_dispatcher['table'] = "start_table" |
|
|
|
|
|
def end_table(self, name, qname): |
|
|
pass |
|
|
end_dispatcher['table'] = "end_table" |
|
562 |
|
|
563 |
def start_labellayer(self, name, qname, attrs): |
def start_labellayer(self, name, qname, attrs): |
564 |
self.aLayer = self.aMap.LabelLayer() |
self.aLayer = self.aMap.LabelLayer() |
|
start_dispatcher['labellayer'] = "start_labellayer" |
|
565 |
|
|
566 |
def start_label(self, name, qname, attrs): |
def start_label(self, name, qname, attrs): |
567 |
x = float(attrs[(None, 'x')]) |
attrs = self.check_attrs(name, attrs, |
568 |
y = float(attrs[(None, 'y')]) |
[AttrDesc("x", True, conversion = float), |
569 |
text = attrs[(None, 'text')] |
AttrDesc("y", True, conversion = float), |
570 |
halign = attrs[(None, 'halign')] |
AttrDesc("text", True), |
571 |
valign = attrs[(None, 'valign')] |
AttrDesc("halign", True, |
572 |
|
conversion = "ascii"), |
573 |
|
AttrDesc("valign", True, |
574 |
|
conversion = "ascii")]) |
575 |
|
x = attrs['x'] |
576 |
|
y = attrs['y'] |
577 |
|
text = attrs['text'] |
578 |
|
halign = attrs['halign'] |
579 |
|
valign = attrs['valign'] |
580 |
|
if halign not in ("left", "center", "right"): |
581 |
|
raise LoadError("Unsupported halign value %r" % halign) |
582 |
|
if valign not in ("top", "center", "bottom"): |
583 |
|
raise LoadError("Unsupported valign value %r" % valign) |
584 |
self.aLayer.AddLabel(x, y, text, halign = halign, valign = valign) |
self.aLayer.AddLabel(x, y, text, halign = halign, valign = valign) |
|
start_dispatcher['label'] = "start_label" |
|
585 |
|
|
586 |
def characters(self, chars): |
def characters(self, chars): |
587 |
pass |
pass |
588 |
|
|
589 |
|
|
590 |
def load_session(filename): |
def load_session(filename, db_connection_callback = None): |
591 |
"""Load a Thuban session from the file object file""" |
"""Load a Thuban session from the file object file |
592 |
dir = os.path.dirname(filename) |
|
593 |
file = open(filename) |
The db_connection_callback, if given should be a callable object |
594 |
handler = ProcessSession(dir) |
that can be called like this: |
595 |
|
db_connection_callback(params, message) |
596 |
parser = make_parser() |
|
597 |
parser.setContentHandler(handler) |
where params is a dictionary containing the known connection |
598 |
parser.setErrorHandler(ErrorHandler()) |
parameters and message is a string with a message why the connection |
599 |
parser.setFeature(xml.sax.handler.feature_namespaces, 1) |
failed. db_connection_callback should return a new dictionary with |
600 |
parser.parse(file) |
corrected and perhaps additional parameters like a password or None |
601 |
|
to indicate that the user cancelled. |
602 |
|
""" |
603 |
|
handler = SessionLoader(db_connection_callback) |
604 |
|
handler.read(filename) |
605 |
|
|
606 |
session = handler.theSession |
session = handler.theSession |
607 |
# Newly loaded session aren't modified |
# Newly loaded session aren't modified |
608 |
session.UnsetModified() |
session.UnsetModified() |
609 |
|
|
610 |
|
handler.Destroy() |
611 |
|
|
612 |
return session |
return session |
613 |
|
|