Parsing a large number of files in different RDF serialisations #1544
Replies: 3 comments
-
Hi @fkraeutli, parallel parsing works only for ntriples not ttl. Or to be more precise ttl that looks like ntriples, so you have a single triple per line with dot in the end, but can use prefixes. We just use ntriples and it works great. |
Beta Was this translation helpful? Give feedback.
-
Ah I didn’t get what the error message meant by ‘well behaved’ turtle files. That’s good to know! In that case, indexing could potentially be sped up by letting rapper output ntriples. |
Beta Was this translation helpful? Give feedback.
-
While I don't disagree that at some point QLever should provide support for a wider range of RDF serializations, I don't see that as a priority. I had a discussion last week with Pierre-Antoine Champin from W3C, he plans to create a Rust RDF community group where we could coordinate work & development on "binary" RDF support front. As in binary executable, suitable for scripting and fast. The use case was converting 60k JSON files in JSON-LD to RDF in a form that it can be ingested by QLever and the main problem is that if this is batched with So the TL;DR statement here is: We should definitely improve native RDF tooling so we have better flexibility in scripting. I use |
Beta Was this translation helpful? Give feedback.
-
I had two issues to solve when indexing RDF data with Qlever:
cat *.ttl
would failI solved this by using the
find
command to concatenate the files and passing the files through the Raptor RDF Utility.The relevant lines in my Qleverfile are:
(I encountered problems with the
parallel-parsing
default so I switched it tofalse
)Beta Was this translation helpful? Give feedback.
All reactions