Bug #47510 core dump in ndbmtd in getting memory on sparc
Submitted: 22 Sep 2009 8:27 Modified: 31 Aug 2010 11:17
Reporter: Jørgen Austvik Email Updates:
Status: Closed Impact on me:
None 
Category:MySQL Cluster: Cluster (NDB) storage engine Severity:S3 (Non-critical)
Version:mysql-5.1-telco-7.0 OS:Solaris (sparc)
Assigned to: Jonas Oreland CPU Architecture:Any

File: Maximum allowed size is 50MB.
Description:
Privacy:

If the data you need to attach is more than 50MB, you should create a compressed archive of the data, split it to 50MB chunks, and upload each of them as a separate attachment.

To split a large file: