I am about to embark on a data processing software project that will require a
LOT of memory (about, ohhh, 6GB or so), and I was wondering if there are any
limitations to how one can use very large chunks of memory under Linux.
Specifically, is there anything to prevent me from malloc()ing 6GB of memory,
then accessing that memory as I would any other buffer? FYI, the application
will be buffering a stream of data, then performing a variety of calculations on
large blocks of data at a time, before writing it out to a socket.
I've been eyeing an 8-way Intel box with gobs of memory, but if there are subtle
issues with using that much memory, I need to know now.
I haven't seen this specifcally addressed, so I figured I should ask you folk.
Any insights/comments/reccomendations would be greatly appreciated.
-- Matthew Nelson Dynamics Technology, Inc. 21311 Hawthorne Blvd., Suite 300, Torrance, CA 90503-5610 Voice: (310) 543-5433 FAX: (310) 543-2117 Email: firstname.lastname@example.org
- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to email@example.com More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
This archive was generated by hypermail 2b29 : Fri Jun 15 2001 - 21:00:14 EST