This page (revision-31) was last changed on 07-Dec-2016 14:14 by HuiLi

This page was created on 18-Sep-2007 12:20 by JianSun

Only authorized users are allowed to rename pages.

Only authorized users are allowed to delete pages.

Page revision history

Version Date Modified Size Author Changes ... Change note
31 07-Dec-2016 14:14 4 KB HuiLi to previous
30 02-Oct-2007 14:39 4 KB HuiLi to previous | to last
29 02-Oct-2007 14:33 4 KB HuiLi to previous | to last
28 02-Oct-2007 14:28 4 KB HuiLi to previous | to last
27 02-Oct-2007 13:46 3 KB HuiLi to previous | to last
26 28-Sep-2007 19:29 2 KB KenDere to previous | to last Comment by KenDere
25 21-Sep-2007 16:49 2 KB JianSun to previous | to last
24 21-Sep-2007 16:40 2 KB JianSun to previous | to last
23 19-Sep-2007 06:42 2 KB David R Williams to previous | to last Comment by David R Williams
22 18-Sep-2007 14:46 1 KB JianSun to previous | to last
21 18-Sep-2007 14:45 1 KB JianSun to previous | to last

Page References

Incoming links Outgoing links

Version management

Difference between version and

At line 1 added 2 lines
[{ALLOW edit EISMainUsers}]
[{ALLOW view Anonymous}]
At line 42 added 44 lines
----
Hi, Jian
This is starting to become a big problem, and I think you're asking the right questions.
However, I don't think we should stop taking large data because the current software doesn't allow us to analyse it. This just means we need to think about . We want to take the best data we can and worry about analysing it later. You never know when an instrument/satellite can fail.
This sounds like a major issue for the team meetings in October.
In the meantime, if anyone has comments, please start making them! This is something that really needs to be sorted out.
--[David R Williams|http://www.mssl.ucl.ac.uk/~drw/], 19-Sep-2007
----
Hi David,
I agree with you and Harry that we should get science data first and then work out the solution, if there is problem to analyse it. So probably we need some modifications on "eis_prep" to improve the memory performance for processing these big fits files.
[JianSun] [{CurrentTimePlugin}]
----
The Hierarchical Data Format (HDF) was made for large files. There are a lot of good tools in IDL for working with HDF files.
--[KenDere|http://null], 28-Sep-2007
----
Hi, Jian,
This problem arises from the limited memory size that IDL can allocate. IDL running on a 32-bit computer can theoretically allocate memory of about 2.3GB (from IDL help files). However, according to my experience, it can only allocate memory of a little more than 1GB, which is not enough to process such a large EIS file with EIS_PREP.
Nevertheless, this problem can be easily solved on a 64-bit computer which running the corresponding version of IDL (also 64-bit). Of course, the computer should have large enough physical memory (maybe 4 GB or more). In this case, IDL can allocate large enough memory to run EIS_PREP to process the EIS data of large size. I have tested it using the file eis_l0_20070824_001531.fits.gz. I processed it with EIS_PREP successfully.
Another way, I guess, to solve this problem is to optimize the IDL code EIS_PREP and related ones, in order to reduce the memory requirement. This may be done by freeing the allocated memory in the code immediately after using it and reducing temporary variables if any. I am thinking about this because I met the same problem when I processed several EIS files with smaller size one by one successively. Therefore, I guess that when we call the code, the allocated memory in the run of the code is not freed after the call.
--[Hui Li|http://solar.pmo.ac.cn/lihui/] 02-Oct-2007
----