[Wrf-users] Fwd: Wrf-users Digest, Vol 138, Issue 16 - (no subject) (afwande juliet)

wrfhelp wrfhelp at ucar.edu
Fri Feb 26 16:40:06 MST 2016


Afwande,

I am not sure whether you have turned on the bucket scheme? If so, then 
you should calculate the total convective and resolvable-scale 
precipitation as below:

RAINC=RAINC+I_RAINC*bucket_mm
RAINNC=RAINNC+I_RAINNC*bucket_mm

Daily and monthly total precipitation should be derived from the total 
precipitation. For example, if you have run WRF for two month, and you 
need to calculate monthly precipitation for the second month. Then you 
need to calculate accumulative precipitation at the end of the first 
month (suppose it is PRECIP1) and that at the end of second month 
(suppose it is PRECIP2). Monthly precipitation for the second month 
should be:

PRECIP2 - PRECIP1

WRFHELP



On 2/26/16 10:50 AM, Jimy Dudhia wrote:
> Maybe they have the bucket scheme on? If so, point them to the user 
> guide on the use of the bucket_mm variables like i_rainnc
> Jimy
>
> ---------- Forwarded message ----------
> From: *afwande juliet* <afwandej965 at gmail.com 
> <mailto:afwandej965 at gmail.com>>
> Date: Fri, Feb 26, 2016 at 1:38 AM
> Subject: Re: [Wrf-users] Wrf-users Digest, Vol 138, Issue 16 - (no 
> subject) (afwande juliet)
> To: Felipe Neris <felipenc2 at gmail.com <mailto:felipenc2 at gmail.com>>
> Cc: wrf users group <wrf-users at ucar.edu <mailto:wrf-users at ucar.edu>>
>
>
> ok thanks
> when I do investigation about my rainc&rainnc, i find that rainc 
> doesnt look cumulative (values decrease and increase randomly within 
> the time of simulation). Looking at rainc further, there are big 
> values upto 1000mm with some negative values in between. However 
> rainnc, looks cumulative and the values increase with time and there 
> are no negative values. Any other variables like temp etc looks 
> reasonable.
>
> And you know that precip=rainc+rainnc
> What could be the problem and how can i correct this?
>
> On Fri, Feb 26, 2016 at 3:16 AM, Felipe Neris <felipenc2 at gmail.com 
> <mailto:felipenc2 at gmail.com>> wrote:
>
>     Hi Juliet,
>     If I got your question right, I suppose the answer is: WRF sums
>     all precipitation that occurs since the beginning of the
>     simulation into the variables RAINNC (from microphysics
>     parameterization) and RAINC (from cumulus parameterization).
>     Therefor, if you want to have the accumulated precipitation of a
>     certain day, you must specify/set the corresponding time for these
>     variables and display.
>     Hope I could be of any help!
>
>     Felipe Neris
>
>     On Thu, Feb 25, 2016 at 3:45 PM, <wrf-users-request at ucar.edu
>     <mailto:wrf-users-request at ucar.edu>> wrote:
>
>         Send Wrf-users mailing list submissions to
>         wrf-users at ucar.edu <mailto:wrf-users at ucar.edu>
>
>         To subscribe or unsubscribe via the World Wide Web, visit
>         http://mailman.ucar.edu/mailman/listinfo/wrf-users
>         or, via email, send a message with subject or body 'help' to
>         wrf-users-request at ucar.edu <mailto:wrf-users-request at ucar.edu>
>
>         You can reach the person managing the list at
>         wrf-users-owner at ucar.edu <mailto:wrf-users-owner at ucar.edu>
>
>         When replying, please edit your Subject line so it is more
>         specific
>         than "Re: Contents of Wrf-users digest..."
>
>
>         Today's Topics:
>
>            1. (no subject) (afwande juliet)
>            2. Nesting and Domain Decomposition (Douglas Lowe)
>            3. Re: Nesting and Domain Decomposition (Tabish Ansari)
>
>
>         ----------------------------------------------------------------------
>
>         Message: 1
>         Date: Thu, 25 Feb 2016 15:45:08 +0300
>         From: afwande juliet <afwandej965 at gmail.com
>         <mailto:afwandej965 at gmail.com>>
>         Subject: [Wrf-users] (no subject)
>         To: wrf users group <wrf-users at ucar.edu
>         <mailto:wrf-users at ucar.edu>>, wrfhelp <wrfhelp at ucar.edu
>         <mailto:wrfhelp at ucar.edu>>
>         Message-ID:
>                
>         <CANVsOojVG3LzL1r82T0oKnBRAU-Nq2vVqHc5GKz4SSj9j0P3Cw at mail.gmail.com
>         <mailto:CANVsOojVG3LzL1r82T0oKnBRAU-Nq2vVqHc5GKz4SSj9j0P3Cw at mail.gmail.com>>
>         Content-Type: text/plain; charset="utf-8"
>
>         I ask this again
>         I have WRF simulations for 1981. The model output is 3hourly,
>         i.e 8
>         timestep in a day
>         When I want daily values or monthly values, do i take every
>         8th timestep to
>         be rain totals for each day and sum them up to get monthly totals?
>         Do I have to divide the units *mm* by any number to get mm/day ?
>
>
>
>         thanks
>         -------------- next part --------------
>         An HTML attachment was scrubbed...
>         URL:
>         http://mailman.ucar.edu/pipermail/wrf-users/attachments/20160225/c10344ce/attachment-0001.html
>
>         ------------------------------
>
>         Message: 2
>         Date: Thu, 25 Feb 2016 13:59:29 +0000
>         From: Douglas Lowe <Douglas.Lowe at manchester.ac.uk
>         <mailto:Douglas.Lowe at manchester.ac.uk>>
>         Subject: [Wrf-users] Nesting and Domain Decomposition
>         To: "wrf-users at ucar.edu <mailto:wrf-users at ucar.edu>"
>         <wrf-users at ucar.edu <mailto:wrf-users at ucar.edu>>
>         Message-ID:
>                
>         <43E6B083008E774B87C4283E0FFA4E70012C7A0C91 at MBXP02.ds.man.ac.uk <mailto:43E6B083008E774B87C4283E0FFA4E70012C7A0C91 at MBXP02.ds.man.ac.uk>>
>         Content-Type: text/plain; charset="us-ascii"
>
>         Hi all,
>
>         I'm running WRF-Chem with a nest of 3 domains, with the
>         settings listed below. I'd like to be
>         able to split this across as many processes as possible in
>         order to speed things up (currently
>         I'm only managing 3x real time, which isn't very good when
>         running multiday simulations).
>         Unfortunately I am finding that WRF hangs when calling the
>         photolysis driver for my 2nd domain
>         (which is the smallest of the domains) if I use too many
>         processors.
>
>         The (relevant) model domain settings are:
>         max_dom                         = 3,
>         e_we                                = 134,  81,   91,
>         e_sn                                = 146,  81,   91,
>         e_vert                              = 41,    41,  41,
>         num_metgrid_levels        = 38,
>         dx                                    = 15000,3000,1000,
>         dy                                    = 15000,3000,1000,
>
>         WRF will run when I split over upto 168 processes (7 nodes on
>         the ARCHER supercomputer),
>         but wont work if I split over 192 (or more) processes (8 nodes
>         on ARCHER).
>
>         Looking at the log messages I *think* that WRF is splitting
>         each domain into the same
>         number of patches, and sending one patch from each domain to a
>         single process for
>         analysis. However, this means that I am limited by the
>         smallest domain as to how many
>         patches I can split a domain into before we end up with
>         patches which are dwarved by
>         the halos around them.
>
>         Would it not make more sense to be able to split each domain
>         into different numbers
>         of patches (so that each patch is of a similar size,
>         regardless of which domain it is from) and
>         send one patch from one domain to a single process (or,
>         perhaps, send more patches from the
>         outer domains to a single process, if needed for balancing
>         computational demands)? And
>         is there anyway for me to do this with WRF?
>
>         Thanks,
>         Doug
>
>         ------------------------------
>
>         Message: 3
>         Date: Thu, 25 Feb 2016 16:37:26 +0000
>         From: Tabish Ansari <tabishumaransari at gmail.com
>         <mailto:tabishumaransari at gmail.com>>
>         Subject: Re: [Wrf-users] Nesting and Domain Decomposition
>         To: Douglas Lowe <Douglas.Lowe at manchester.ac.uk
>         <mailto:Douglas.Lowe at manchester.ac.uk>>
>         Cc: "wrf-users at ucar.edu <mailto:wrf-users at ucar.edu>"
>         <wrf-users at ucar.edu <mailto:wrf-users at ucar.edu>>
>         Message-ID:
>                
>         <CALLVTyvwh3nimJczxjfAy+gNML1PfJpDudJT8TQhkDQrCLnwZw at mail.gmail.com
>         <mailto:CALLVTyvwh3nimJczxjfAy%2BgNML1PfJpDudJT8TQhkDQrCLnwZw at mail.gmail.com>>
>         Content-Type: text/plain; charset="utf-8"
>
>         Hi Doug,
>
>         I'm not too knowledgeable in this but have some literature
>         which might be
>         of relevance. Please have a look at the attached files.
>
>         Cheers,
>
>         Tabish
>
>         Tabish U Ansari
>         PhD student, Lancaster Environment Center
>         Lancaster Univeristy
>         Bailrigg, Lancaster,
>         LA1 4YW, United Kingdom
>
>         On 25 February 2016 at 13:59, Douglas Lowe
>         <Douglas.Lowe at manchester.ac.uk
>         <mailto:Douglas.Lowe at manchester.ac.uk>>
>         wrote:
>
>         > Hi all,
>         >
>         > I'm running WRF-Chem with a nest of 3 domains, with the
>         settings listed
>         > below. I'd like to be
>         > able to split this across as many processes as possible in
>         order to speed
>         > things up (currently
>         > I'm only managing 3x real time, which isn't very good when
>         running
>         > multiday simulations).
>         > Unfortunately I am finding that WRF hangs when calling the
>         photolysis
>         > driver for my 2nd domain
>         > (which is the smallest of the domains) if I use too many
>         processors.
>         >
>         > The (relevant) model domain settings are:
>         > max_dom                         = 3,
>         > e_we = 134,  81,   91,
>         > e_sn = 146,  81,   91,
>         > e_vert = 41,    41,  41,
>         > num_metgrid_levels        = 38,
>         > dx   = 15000,3000,1000,
>         > dy   = 15000,3000,1000,
>         >
>         > WRF will run when I split over upto 168 processes (7 nodes
>         on the ARCHER
>         > supercomputer),
>         > but wont work if I split over 192 (or more) processes (8
>         nodes on ARCHER).
>         >
>         > Looking at the log messages I *think* that WRF is splitting
>         each domain
>         > into the same
>         > number of patches, and sending one patch from each domain to
>         a single
>         > process for
>         > analysis. However, this means that I am limited by the
>         smallest domain as
>         > to how many
>         > patches I can split a domain into before we end up with
>         patches which are
>         > dwarved by
>         > the halos around them.
>         >
>         > Would it not make more sense to be able to split each domain
>         into
>         > different numbers
>         > of patches (so that each patch is of a similar size,
>         regardless of which
>         > domain it is from) and
>         > send one patch from one domain to a single process (or,
>         perhaps, send more
>         > patches from the
>         > outer domains to a single process, if needed for balancing
>         computational
>         > demands)? And
>         > is there anyway for me to do this with WRF?
>         >
>         > Thanks,
>         > Doug
>         > _______________________________________________
>         > Wrf-users mailing list
>         > Wrf-users at ucar.edu <mailto:Wrf-users at ucar.edu>
>         > http://mailman.ucar.edu/mailman/listinfo/wrf-users
>         >
>         -------------- next part --------------
>         An HTML attachment was scrubbed...
>         URL:
>         http://mailman.ucar.edu/pipermail/wrf-users/attachments/20160225/02afb9aa/attachment.html
>         -------------- next part --------------
>         A non-text attachment was scrubbed...
>         Name: WRF-HPC.pdf
>         Type: application/pdf
>         Size: 243897 bytes
>         Desc: not available
>         Url :
>         http://mailman.ucar.edu/pipermail/wrf-users/attachments/20160225/02afb9aa/attachment.pdf
>         -------------- next part --------------
>         A non-text attachment was scrubbed...
>         Name: WRF-chapter-multicore.pdf
>         Type: application/pdf
>         Size: 230093 bytes
>         Desc: not available
>         Url :
>         http://mailman.ucar.edu/pipermail/wrf-users/attachments/20160225/02afb9aa/attachment-0001.pdf
>         -------------- next part --------------
>         A non-text attachment was scrubbed...
>         Name: CUDA-WRF_ppt.pdf
>         Type: application/pdf
>         Size: 2314206 bytes
>         Desc: not available
>         Url :
>         http://mailman.ucar.edu/pipermail/wrf-users/attachments/20160225/02afb9aa/attachment-0002.pdf
>
>         ------------------------------
>
>         _______________________________________________
>         Wrf-users mailing list
>         Wrf-users at ucar.edu <mailto:Wrf-users at ucar.edu>
>         http://mailman.ucar.edu/mailman/listinfo/wrf-users
>
>
>         End of Wrf-users Digest, Vol 138, Issue 16
>         ******************************************
>
>
>
>     _______________________________________________
>     Wrf-users mailing list
>     Wrf-users at ucar.edu <mailto:Wrf-users at ucar.edu>
>     http://mailman.ucar.edu/mailman/listinfo/wrf-users
>
>
>
> _______________________________________________
> Wrf-users mailing list
> Wrf-users at ucar.edu <mailto:Wrf-users at ucar.edu>
> http://mailman.ucar.edu/mailman/listinfo/wrf-users
>
>

-------------- next part --------------
An HTML attachment was scrubbed...
URL: http://mailman.ucar.edu/pipermail/wrf-users/attachments/20160226/01c83c3b/attachment-0001.html 


More information about the Wrf-users mailing list