Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Possible memory leak in SPARQL endpoint? #123

Open
cKlee opened this issue Apr 18, 2016 · 1 comment
Open

Possible memory leak in SPARQL endpoint? #123

cKlee opened this issue Apr 18, 2016 · 1 comment
Assignees

Comments

@cKlee
Copy link

cKlee commented Apr 18, 2016

Hi,
I'm using the solrdf-1.0 branch with JRE 1.7.0 and successfully loaded aprox. 125000000 documents to the store (and optimized). Solr queries are fine. But with simple SPARQL queries I always get a "java heap space out of memory" exception (even with LIMIT 1).

I'm running a 4 CPU 64 bit Debian with 16G RAM. I'm not very familiar with Maven, but I tried to give Solrdf a little bit more RAM with

$ mvn -DargLine="-Xmx8g" cargo:run

... but with the same exception.

@agazzarini
Copy link
Member

Hi @cKlee,
unfortunately this is part of #96 which is a huge work, still pending. The current implementation of SolRDF uses the general-purpose SPARQL Algebra implementation bundled with Jena, which is good for small or in-memory datasets.

A more sophisticated logic is needed here, in order to gain advantage from the underlying inverted index and most important, to avoid what you're seeing.

Specifically, even small queries containing joins or count() with simple graph patterns, completely scans the entire index; I guess that's the underlying reason of your out of memory issue.

The bad thing is that I have no a precise idea about when this thing will be done, as it is complicated and it is taking me a lot of time.

AG

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants