[Wrf-users] Wrf-users Digest, Vol 138, Issue 16 - (no subject) (afwande juliet)

Felipe Neris felipenc2 at gmail.com
Thu Feb 25 17:16:12 MST 2016


Hi Juliet,
If I got your question right, I suppose the answer is: WRF sums all
precipitation that occurs since the beginning of the simulation into the
variables RAINNC (from microphysics parameterization) and RAINC (from
cumulus parameterization). Therefor, if you want to have the accumulated
precipitation of a certain day, you must specify/set the corresponding time
for these variables and display.
Hope I could be of any help!

Felipe Neris


On Thu, Feb 25, 2016 at 3:45 PM, <wrf-users-request at ucar.edu> wrote:

> Send Wrf-users mailing list submissions to
>         wrf-users at ucar.edu
>
> To subscribe or unsubscribe via the World Wide Web, visit
>         http://mailman.ucar.edu/mailman/listinfo/wrf-users
> or, via email, send a message with subject or body 'help' to
>         wrf-users-request at ucar.edu
>
> You can reach the person managing the list at
>         wrf-users-owner at ucar.edu
>
> When replying, please edit your Subject line so it is more specific
> than "Re: Contents of Wrf-users digest..."
>
>
> Today's Topics:
>
>    1. (no subject) (afwande juliet)
>    2. Nesting and Domain Decomposition (Douglas Lowe)
>    3. Re: Nesting and Domain Decomposition (Tabish Ansari)
>
>
> ----------------------------------------------------------------------
>
> Message: 1
> Date: Thu, 25 Feb 2016 15:45:08 +0300
> From: afwande juliet <afwandej965 at gmail.com>
> Subject: [Wrf-users] (no subject)
> To: wrf users group <wrf-users at ucar.edu>, wrfhelp <wrfhelp at ucar.edu>
> Message-ID:
>         <
> CANVsOojVG3LzL1r82T0oKnBRAU-Nq2vVqHc5GKz4SSj9j0P3Cw at mail.gmail.com>
> Content-Type: text/plain; charset="utf-8"
>
> I ask this again
> I have WRF simulations for 1981. The model output is 3hourly, i.e 8
> timestep in a day
> When I want daily values or monthly values, do i take every 8th timestep to
> be rain totals for each day and sum them up to get monthly totals?
> Do I have to divide the units *mm* by any number to get mm/day ?
>
>
>
> thanks
> -------------- next part --------------
> An HTML attachment was scrubbed...
> URL:
> http://mailman.ucar.edu/pipermail/wrf-users/attachments/20160225/c10344ce/attachment-0001.html
>
> ------------------------------
>
> Message: 2
> Date: Thu, 25 Feb 2016 13:59:29 +0000
> From: Douglas Lowe <Douglas.Lowe at manchester.ac.uk>
> Subject: [Wrf-users] Nesting and Domain Decomposition
> To: "wrf-users at ucar.edu" <wrf-users at ucar.edu>
> Message-ID:
>         <43E6B083008E774B87C4283E0FFA4E70012C7A0C91 at MBXP02.ds.man.ac.uk>
> Content-Type: text/plain; charset="us-ascii"
>
> Hi all,
>
> I'm running WRF-Chem with a nest of 3 domains, with the settings listed
> below. I'd like to be
> able to split this across as many processes as possible in order to speed
> things up (currently
> I'm only managing 3x real time, which isn't very good when running
> multiday simulations).
> Unfortunately I am finding that WRF hangs when calling the photolysis
> driver for my 2nd domain
> (which is the smallest of the domains) if I use too many processors.
>
> The (relevant) model domain settings are:
> max_dom                         = 3,
> e_we                                = 134,  81,   91,
> e_sn                                = 146,  81,   91,
> e_vert                              = 41,    41,  41,
> num_metgrid_levels        = 38,
> dx                                    = 15000,3000,1000,
> dy                                    = 15000,3000,1000,
>
> WRF will run when I split over upto 168 processes (7 nodes on the ARCHER
> supercomputer),
> but wont work if I split over 192 (or more) processes (8 nodes on ARCHER).
>
> Looking at the log messages I *think* that WRF is splitting each domain
> into the same
> number of patches, and sending one patch from each domain to a single
> process for
> analysis. However, this means that I am limited by the smallest domain as
> to how many
> patches I can split a domain into before we end up with patches which are
> dwarved by
> the halos around them.
>
> Would it not make more sense to be able to split each domain into
> different numbers
> of patches (so that each patch is of a similar size, regardless of which
> domain it is from) and
> send one patch from one domain to a single process (or, perhaps, send more
> patches from the
> outer domains to a single process, if needed for balancing computational
> demands)? And
> is there anyway for me to do this with WRF?
>
> Thanks,
> Doug
>
> ------------------------------
>
> Message: 3
> Date: Thu, 25 Feb 2016 16:37:26 +0000
> From: Tabish Ansari <tabishumaransari at gmail.com>
> Subject: Re: [Wrf-users] Nesting and Domain Decomposition
> To: Douglas Lowe <Douglas.Lowe at manchester.ac.uk>
> Cc: "wrf-users at ucar.edu" <wrf-users at ucar.edu>
> Message-ID:
>         <
> CALLVTyvwh3nimJczxjfAy+gNML1PfJpDudJT8TQhkDQrCLnwZw at mail.gmail.com>
> Content-Type: text/plain; charset="utf-8"
>
> Hi Doug,
>
> I'm not too knowledgeable in this but have some literature which might be
> of relevance. Please have a look at the attached files.
>
> Cheers,
>
> Tabish
>
> Tabish U Ansari
> PhD student, Lancaster Environment Center
> Lancaster Univeristy
> Bailrigg, Lancaster,
> LA1 4YW, United Kingdom
>
> On 25 February 2016 at 13:59, Douglas Lowe <Douglas.Lowe at manchester.ac.uk>
> wrote:
>
> > Hi all,
> >
> > I'm running WRF-Chem with a nest of 3 domains, with the settings listed
> > below. I'd like to be
> > able to split this across as many processes as possible in order to speed
> > things up (currently
> > I'm only managing 3x real time, which isn't very good when running
> > multiday simulations).
> > Unfortunately I am finding that WRF hangs when calling the photolysis
> > driver for my 2nd domain
> > (which is the smallest of the domains) if I use too many processors.
> >
> > The (relevant) model domain settings are:
> > max_dom                         = 3,
> > e_we                                = 134,  81,   91,
> > e_sn                                = 146,  81,   91,
> > e_vert                              = 41,    41,  41,
> > num_metgrid_levels        = 38,
> > dx                                    = 15000,3000,1000,
> > dy                                    = 15000,3000,1000,
> >
> > WRF will run when I split over upto 168 processes (7 nodes on the ARCHER
> > supercomputer),
> > but wont work if I split over 192 (or more) processes (8 nodes on
> ARCHER).
> >
> > Looking at the log messages I *think* that WRF is splitting each domain
> > into the same
> > number of patches, and sending one patch from each domain to a single
> > process for
> > analysis. However, this means that I am limited by the smallest domain as
> > to how many
> > patches I can split a domain into before we end up with patches which are
> > dwarved by
> > the halos around them.
> >
> > Would it not make more sense to be able to split each domain into
> > different numbers
> > of patches (so that each patch is of a similar size, regardless of which
> > domain it is from) and
> > send one patch from one domain to a single process (or, perhaps, send
> more
> > patches from the
> > outer domains to a single process, if needed for balancing computational
> > demands)? And
> > is there anyway for me to do this with WRF?
> >
> > Thanks,
> > Doug
> > _______________________________________________
> > Wrf-users mailing list
> > Wrf-users at ucar.edu
> > http://mailman.ucar.edu/mailman/listinfo/wrf-users
> >
> -------------- next part --------------
> An HTML attachment was scrubbed...
> URL:
> http://mailman.ucar.edu/pipermail/wrf-users/attachments/20160225/02afb9aa/attachment.html
> -------------- next part --------------
> A non-text attachment was scrubbed...
> Name: WRF-HPC.pdf
> Type: application/pdf
> Size: 243897 bytes
> Desc: not available
> Url :
> http://mailman.ucar.edu/pipermail/wrf-users/attachments/20160225/02afb9aa/attachment.pdf
> -------------- next part --------------
> A non-text attachment was scrubbed...
> Name: WRF-chapter-multicore.pdf
> Type: application/pdf
> Size: 230093 bytes
> Desc: not available
> Url :
> http://mailman.ucar.edu/pipermail/wrf-users/attachments/20160225/02afb9aa/attachment-0001.pdf
> -------------- next part --------------
> A non-text attachment was scrubbed...
> Name: CUDA-WRF_ppt.pdf
> Type: application/pdf
> Size: 2314206 bytes
> Desc: not available
> Url :
> http://mailman.ucar.edu/pipermail/wrf-users/attachments/20160225/02afb9aa/attachment-0002.pdf
>
> ------------------------------
>
> _______________________________________________
> Wrf-users mailing list
> Wrf-users at ucar.edu
> http://mailman.ucar.edu/mailman/listinfo/wrf-users
>
>
> End of Wrf-users Digest, Vol 138, Issue 16
> ******************************************
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: http://mailman.ucar.edu/pipermail/wrf-users/attachments/20160225/64dde822/attachment-0001.html 


More information about the Wrf-users mailing list