[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
Thredds out of memory
- Subject: Thredds out of memory
- Date: Thu, 24 Mar 2005 15:07:49 +1100
Hi guys,
I am looking at serving some very large files through thredds. I had
found through trial-and-error that on one particular server, somewhere
between 60Mb and 300Mb thredds stopped being able to start serving up
files before the client timed out.
Unfortunately, this machine services a number of people so I had to do
my testing elsewhere. I have a 579Mb NetCDF file on my desktop machine,
and tried doing a local test with this, installing my file server and
the thredds server on it. What I found was that the thredds server was
running out of heap space. Now, I know I can alter the amount of heap
space the JVM has available somehow, and that's what I'll try next, but
I don't know whether that's a reliable solution. I don't really know how
much memory thredds needs on top of the size of the file it's trying to
serve, and of course multiple incoming requests might also affect this -
I don't know how tomcat deals with that kind of thing in terms of
creating new JVM instances etc.
Here is the error from catalina.out:
DODServlet ERROR (anyExceptionHandler): java.lang.OutOfMemoryError: Java
heap space
requestState:
dataset: 'verylarge.nc'
suffix: 'dods'
CE: ''
compressOK: false
InitParameters:
maxAggDatasetsCached: '20'
maxNetcdfFilesCached: '100'
maxDODSDatasetsCached: '100'
displayName: 'THREDDS/DODS Aggregation/NetCDF/Catalog Server'
java.lang.OutOfMemoryError: Java heap space
So my question is: what's the best way to make a reliable server than
can serve these large files?
Cheers,
-Tennessee