On Wed, May 13, 2009 at 10:50 PM, Glenn Tarbox, PhD <gl...@tarbox.org>wrote:
> I'm using the latest version of Sage (3.4.2) which is python 2.5 and numpy > something or other (I will do more digging presently) > > I'm able to map large files and access all the elements unless I'm using > slices > > so, for example: > > fp = np.memmap("/mnt/hdd/data/mmap/numpy1e10.mmap", dtype='float64', > mode='r+', shape=(10000000000,)) > > which is 1e10 doubles if you don't wanna count the zeros > > gives full access to a 75 GB memory image > > But when I do: > > fp[:] = 1.0 > np.sum(fp) > > I get 1410065408.0 as the result > As doubles, that is more than 2**33 bytes, so I expect there is something else going on. How much physical memory/swap memory do you have? This could also be a python problem since python does the memmap. Chuck
_______________________________________________ Numpy-discussion mailing list Numpy-discussion@scipy.org http://mail.scipy.org/mailman/listinfo/numpy-discussion