c语言MPI并行程序出现错误:
[cb81ea2c449b:09101] *** Process received signal ***
[cb81ea2c449b:09101] Signal: Segmentation fault (11)
[cb81ea2c449b:09101] Signal code: (128)
[cb81ea2c449b:09101] Failing at address: (nil)
[cb81ea2c449b:09104] *** Process received signal ***
[cb81ea2c449b:09104] Signal: Segmentation fault (11)
[cb81ea2c449b:09104] Signal code: (128)
[cb81ea2c449b:09104] Failing at address: (nil)
[cb81ea2c449b:09104] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x110c0)[0x7fe2b8dba0c0]
[cb81ea2c449b:09104] [ 1] /lib/x86_64-linux-gnu/libc.so.6(+0x120e6c)[0x7fe2b8b2ae6c]
[cb81ea2c449b:09104] [ 2] [cb81ea2c449b:09101] [ 0] /lib/x86_64-linux-gnu/libpthread.so.0(+0x110c0)[0x7f5d5708e0c0]
[cb81ea2c449b:09101] [ 1] /lib/x86_64-linux-gnu/libc.so.6(+0x120e6c)[0x7f5d56dfee6c]
[cb81ea2c449b:09101] [ 2] /usr/local/lib/libopen-pal.so.40(opal_unsetenv+0x6c)[0x7f5d56781ccc]
[cb81ea2c449b:09101] [ 3] /usr/local/lib/openmpi/mca_schizo_orte.so(+0xab8)[0x7f5d551e7ab8]
[cb81ea2c449b:09101] [ 4] /usr/local/lib/libopen-rte.so.40(orte_schizo_base_finalize+0x2f)[0x7f5d56aa7b5f]
[cb81ea2c449b:09101] [ 5] /usr/local/lib/libopen-rte.so.40(orte_finalize+0x95)[0x7f5d56a3dc85]
[cb81ea2c449b:09101] [ 6] /usr/local/lib/libmpi.so.40(ompi_mpi_finalize+0x872)[0x7f5d572e5562]
[cb81ea2c449b:09101] [ 7] /usr/local/lib/libopen-pal.so.40(opal_unsetenv+0x6c)[0x7fe2b84adccc]
[cb81ea2c449b:09104] [ 3] /usr/local/lib/openmpi/mca_schizo_orte.so(+0xab8)[0x7fe2b6f13ab8]
[cb81ea2c449b:09104] [ 4] /usr/local/lib/libopen-rte.so.40(orte_schizo_base_finalize+0x2f)[0x7fe2b87d3b5f]
[cb81ea2c449b:09104] [ 5] /usr/local/lib/libopen-rte.so.40(orte_finalize+0x95)[0x7fe2b8769c85]
[cb81ea2c449b:09104] [ 6] /usr/local/lib/libmpi.so.40(ompi_mpi_finalize+0x872)[0x7fe2b9011562]
[cb81ea2c449b:09104] [ 7] --------------------------------------------------------------------------
Primary job terminated normally, but 1 process returned
a non-zero exit code. Per user-direction, the job has been aborted.
--------------------------------------------------------------------------
--------------------------------------------------------------------------
mpirun noticed that process rank 2 with PID 0 on node cb81ea2c449b exited on signal 11 (Segmentation fault).
mpi c语言代码如下:
#include "mpi.h"
#include<stdio.h>
#include<stdlib.h>
int main()
{
int size,rank,m,n,*recvcounts,*disp,i,j=0,index=0,sum=0;
MPI_Comm mm=MPI_COMM_WORLD;//记得名字不要重名啊
double *vector,*matrix,*la;
MPI_Init(NULL,NULL);
MPI_Comm_size(mm,&size);
MPI_Comm_rank(mm,&rank);
if(rank==0)
{
printf("enter m and n:\n");
scanf("%d %d",&m,&n);
vector=(double*)malloc(sizeof(double)*n);
matrix=(double*)malloc(sizeof(double)*m*n);
printf("enter vector(%d):\n",n);
for(i=0;i<n;i++)
scanf("%lf",&vector[i]);//记得要加&符号
printf("enter matrix:\n");
for(i=0;i<m*n;i++)
scanf("%lf",&matrix[i]);
}
MPI_Bcast(&m,1,MPI_INT,0,mm);
MPI_Bcast(&n,1,MPI_INT,0,mm);
MPI_Bcast(vector,n,MPI_DOUBLE,0,mm);
recvcounts=(int*)malloc(sizeof(int)*size);
disp=(int*)malloc(sizeof(int)*size);
for(i=0;i<size;i++)
{
recvcounts[i]=(m*n)/size;
if(i<(m*n)%size)
recvcounts[i]++;
disp[i]=(i>0?recvcounts[i-1]+disp[i-1]:0);
}
la=(double*)malloc(sizeof(double)*recvcounts[rank]);
MPI_Scatterv(matrix,recvcounts,disp,MPI_DOUBLE,la,recvcounts[rank],MPI_DOUBLE,0,mm);
index=disp[rank];
for(i=0;i<recvcounts[rank];i++)
{
la[i]*=vector[(index+i)%n];
}
MPI_Gatherv(la,recvcounts[rank],MPI_DOUBLE,matrix,recvcounts,disp,MPI_DOUBLE,0,mm);
if(rank==0) //要加
{
for(i=0;i<m*n;i++)
{
printf("%lf ",matrix[i]);
if(i%n==n-1) printf("\n");
}
}
MPI_Finalize();
}
为什么编译运行这个矩阵乘以向量的程序会出现
```c
#include <stdio.h>
#include <stdlib.h>
#include <mpi.h>
#define N 4
int main(int argc, char *argv[]) {
int rank, size;
int i, j;
int mat[N][N], vec[N], result[N], recvbuf[N];
MPI_Status status;
MPI_Init(&argc, &argv);
MPI_Comm_rank(MPI_COMM_WORLD, &rank);
MPI_Comm_size(MPI_COMM_WORLD, &size);
if (rank == 0) {
// 进程0读取矩阵和向量
printf("Enter the matrix:\n");
for (i = 0; i < N; i++) {
for (j = 0; j < N; j++) {
scanf("%d", &mat[i][j]);
}
}
printf("Enter the vector:\n");
for (i = 0; i < N; i++) {
scanf("%d", &vec[i]);
}
// 将任务分配到各个进程
for (i = 1; i < size; i++) {
MPI_Send(mat[i-1], N, MPI_INT, i, 0, MPI_COMM_WORLD);
}
MPI_Bcast(vec, N, MPI_INT, 0, MPI_COMM_WORLD);
// 进程0计算自己的部分
for (i = 0; i < N; i++) {
result[i] = 0;
for (j = 0; j < N; j++) {
result[i] += mat[0][j] * vec[j];
}
}
// 接收其他进程的计算结果
for (i = 1; i < size; i++) {
MPI_Recv(recvbuf, N, MPI_INT, i, 0, MPI_COMM_WORLD, &status);
for (j = 0; j < N; j++) {
result[j] += recvbuf[j];
}
}
// 打印计算结果
printf("Result:\n");
for (i = 0; i < N; i++) {
printf("%d ", result[i]);
}
printf("\n");
} else {
// 接收任务
MPI_Recv(mat, N*N, MPI_INT, 0, 0, MPI_COMM_WORLD, &status);
MPI_Bcast(vec, N, MPI_INT, 0, MPI_COMM_WORLD);
// 计算自己的部分
for (i = 0; i < N; i++) {
recvbuf[i] = 0;
for (j = 0; j < N; j++) {
recvbuf[i] += mat[i][j] * vec[j];
}
}
// 发送计算结果
MPI_Send(recvbuf, N, MPI_INT, 0, 0, MPI_COMM_WORLD);
}
MPI_Finalize();
return 0;
}
```