14:56:25 RRSAgent has joined #hcls 14:56:30 logging to https://www.w3.org/2025/05/08-hcls-irc 14:56:40 rrsagent, make logs public 14:56:53 Meeting: FHIR RDF 14:56:57 Chair: David Booth 15:06:14 Topic: Complex Data Types 15:08:05 https://www.lassila.org/publications/2024/HartigEtAl_SPARQLCDTs_PosterPaper2024.pdf 15:08:49 Present: Erich Bremer, Detlef Grittner, David Booth 15:11:14 erich: Complex Data Types (CDT) is implemented in Jena 15:11:46 ... And RDF lists can be turned into CDT arrays. 15:12:09 ... CDT is a proposal 15:12:41 ... We also need a standardbinary representation of RDF. 15:13:03 ... RDF HDT allows query on it. 15:13:42 ... They treat all literals as a string, which doesn't work well for a lot of numeric data, but it's very fast. 15:13:52 ... Great for huge amounts of data. 15:14:16 ... I used it with Apache Arrow that I made. 15:14:50 ... HDF5 is use in the scientific world, like an intelligent zip file. Chunks can be compressed, but you can pull out parts. 15:15:11 ... I'm putting the guts of HDT into HDF5. 15:16:34 ... OpenLink Software won't touch it if it isn't a standard. 15:21:48 ... I harvested all of the imaging commons data and it was 90 TB of data, and pulled out all the metadata in a week. 15:22:39 ... I'm working on a SOLID project -- Java library that wraps a read/write storage for linked web storage. 15:22:48 ... Want to use it for image annotations and displays. 15:25:25 dbooth: Any SPARQL implementatoins of CDT yet? 15:25:42 erich: Only the Jena SPARQLstore. 15:31:07 erich: CDT paper: https://awslabs.github.io/SPARQL-CDTs/spec/latest.html 15:31:17 ... also this: https://www.amazon.science/publications/datatypes-for-lists-and-maps-in-rdf-literals 15:32:11 ... I also suggested allowing JSON paths 15:34:44 dbooth: Community group is a good step toward official W3C working group 15:36:35 detlef: From a theoretical standpoing, the hierarchical approach in DICOM is the right way to do it 15:36:49 ... You have to connect to the UIDs of the entities 15:38:31 https://dl.gi.de/items/6ae82b4a-c2c8-4d7e-b45b-088e82080f99 15:39:16 i/hierar/erich: Need to remove redundancy in the file 15:41:06 erich: The HDT source code is GPL. I'd rather have it MIT or Apache 2 license. 15:41:20 ... but it squeezes the data down a lot. 15:42:05 ... I doesn't need to be serialized/deserialized. Like a direct memory copy. 15:43:21 ... HDF5 is used for a lot of learning models. 15:44:43 ... The ordering of the HDT data is not currently usable. 15:45:05 ... I'm stuffing HDT data into HDF5. 15:59:26 ADJOURNED 15:59:31 rrsagent, draft minutes 15:59:32 I have made the request to generate https://www.w3.org/2025/05/08-hcls-minutes.html dbooth