I think Xpath in postgresql is fast enough. I am dumping the raw xml file into a table and then generating data from it. I want all the data to be consistent thats why I am using a database.
I am planning to use triggers to generate content of table1_master...table2_master, etc...
master -> table 1_master
-> table 2_master
-> table 3_master -> table 1_table3
On Sun, Jan 29, 2017 at 9:45 PM, David G. Johnston <david.g.johnston@xxxxxxxxx> wrote:
On Saturday, January 28, 2017, Rita <rmorgan466@xxxxxxxxx> wrote:After xmltest has been populated, I can run xpath and unest to get my data into a row but I would like to store that result in another table, I am guessing I should look into triggers for something like that?I suspect that using xpath in the database is not the right tool for doing what you describe. Whatever middleware layer receives the XML should be considered for the logic of deserialization and storage to the database in normalized form. If you do want something like that in the database I'd probably write a volatile function the receives xml and does whatever it needs to do. I cannot imagine the indirection of a trigger would be gainful here.In particular, at scale you'd probably be better off with using a streaming parser instead of a DOM one.David J.
--- Get your facts first, then you can distort them as you please.--