diff --git a/Documentation/Examples/Containers/DistributedArrayExample.cpp b/Documentation/Examples/Containers/DistributedArrayExample.cpp
index 8e191f652570771577589cd4920a1d0e3f9100f5..02f926ffae7f9b06c01d56c570de0718f5452430 100644
--- a/Documentation/Examples/Containers/DistributedArrayExample.cpp
+++ b/Documentation/Examples/Containers/DistributedArrayExample.cpp
@@ -17,9 +17,9 @@ void distributedArrayExample()
    using IndexType = typename ArrayType::IndexType;
    using LocalRangeType = typename ArrayType::LocalRangeType;
 
-   const MPI_Comm group = TNL::MPI::AllGroup();
-   //const int rank = TNL::MPI::GetRank(group);
-   const int nproc = TNL::MPI::GetSize(group);
+   const MPI_Comm communicator = MPI_COMM_WORLD;
+   //const int rank = TNL::MPI::GetRank(communicator);
+   const int nproc = TNL::MPI::GetSize(communicator);
 
    /***
     * We set size to prime number to force non-uniform distribution of the distributed array.
@@ -27,9 +27,9 @@ void distributedArrayExample()
    const int size = 97;
    const int ghosts = (nproc > 1) ? 4 : 0;
 
-   const LocalRangeType localRange = Containers::Partitioner< IndexType >::splitRange( size, group );
-   ArrayType a( localRange, ghosts, size, group );
-   a.forElements( 0, size, [=] __cuda_callable__ ( const int idx, int& value ) { value = idx; } );
+   const LocalRangeType localRange = Containers::Partitioner< IndexType >::splitRange( size, communicator );
+   ArrayType a( localRange, ghosts, size, communicator );
+   a.forElements( 0, size, [] __cuda_callable__ ( int idx, int& value ) { value = idx; } );
    //LocalArrayType localArray = a;
    //std::cout << a << std::endl;