<p><b>duda</b> 2012-07-18 15:56:26 -0600 (Wed, 18 Jul 2012)</p><p>BRANCH COMMIT<br>
<br>
Define nVertLevelsSolve after call to mpas_block_creator_finalize_block_init() in input module.<br>
Currently, nVertLevelsSolve = nVertLevels, since we don't do vertical decomposition.<br>
<br>
<br>
M src/framework/mpas_io_input.F<br>
</p><hr noshade><pre><font color="gray">Modified: branches/omp_blocks/multiple_blocks/src/framework/mpas_io_input.F
===================================================================
--- branches/omp_blocks/multiple_blocks/src/framework/mpas_io_input.F        2012-07-18 17:20:39 UTC (rev 2033)
+++ branches/omp_blocks/multiple_blocks/src/framework/mpas_io_input.F        2012-07-18 21:56:26 UTC (rev 2034)
@@ -216,6 +216,8 @@
#include "dim_dummy_args.inc"
, nCellsSolveField, nEdgesSolveField, nVerticesSolveField, indexToCellID_Block, indexToEdgeID_Block, indexToVertexID_Block)
+ domain % blocklist % mesh % nVertLevelsSolve = domain % blocklist % mesh % nVertLevels ! No vertical decomp yet...
+
call mpas_io_input_init(input_obj, domain % blocklist, domain % dminfo)
call MPAS_readStreamAtt(input_obj % io_stream, 'sphere_radius', r_sphere_radius, ierr)
@@ -244,6 +246,7 @@
do while (associated(block_ptr))
block_ptr % mesh % sphere_radius = domain % blocklist % mesh % sphere_radius
block_ptr % mesh % on_a_sphere = domain % blocklist % mesh % on_a_sphere
+ block_ptr % mesh % nVertLevelsSolve = domain % blocklist % mesh % nVertLevelsSolve ! No vertical decomp yet...
! Link the sendList and recvList pointers in each field type to the appropriate lists
! in parinfo, e.g., cellsToSend and cellsToRecv; in future, it can also be extended to
</font>
</pre>