Hi,
Currently I am doing the memory free experiment based on the large-sized matrices, I just found if I increase the size to 10000, the memory free speed is really slow even if I did it in a parallel mode. I wonder if someone else can help me resolve...
Since we are supposed to test with probabilities 0.01, 0.02, and 0.05, and for each one we generate a matrix X and a matrix Y. We compress each of the 3 X matrices and 3 Y matrices, producing 6 B matrices and 6 C matrices.
So which one do we write to...
Well actually it would usually make sense. Pretend we have 1 thread and 1 core. Every time one of the items in a row or column of the matrix needs to be accessed, that data needs to be pulled from the RAM or a cache into the CPU, this can take hundre...
The speed of writing the result matrix B and C to file is extremely slow.
The reason is the result matrix of B and C has column number nearly 100000 so it is very big.
Can I write binary format to the file instead? I've tested it is much faster than w...
Thanks, probably this is the reason why it get killed.
However the tricky part is that if I run it without script, just . project.
It can run sucessfully sometimes.
Sorry, the first line of the previous reply is incomplete
Your computer may also have pagefile.sys or swapfile enabled, so it is possible that you will not encounter an OOM (Out of Memory) error as it will swap memory between physical memory and disk...
I understand that but I dont think virual memeory management would affect the use of memory, this is what operation system and c library should gurantee.
If malloc return sucess I should be able to use it freely. In fact I have run my program on my la...
You can use squeue -u USER to get all of your currently running job, and then use sstat -j job-id to show the details of your job, including an entry representing memory usage.
It's important to note that simply allocating memory via malloc() ...
when I tried to alloc some memory on setonix, with script
bin bash
SBATCH --nodes 1
SBTACH --ntasks 28
SBATCH --partition work
SBATCH --account courses0101
SBATCH --mem 200G
export OMP SCHEDULE "static,3125"
gcc -m64 -fopenmp -o project . project.c
s...
I believe a proper implementation should be able to perform the multiplication for p 0.05 and N 100,000 in under ten minutes. In fact, ten minutes is a generous estimate, and your program will likely complete the task in less time.
Isn't that changing the number of "cores". It seems --cpus-per-task can be up to 128, but I assume we should set it 28 since the task says it should use 28 cores. Then we can change the number of threads used with omp set num threads while keeping co...
Hi
I am struggling to get my sparse matrix multiplication algorithm to any speed close to realistic when raising it above 1000x1000, without any parallel programming.
My algorithm works by going through each value in the multiplying matrix, and then f...
I agree. It might work in later versions, but earlier versions of OpenMP could exhibit undefined behavior.
In this case, I think we might just need to choose the option that makes the most sense.