[Wrf-users] Fwd: Wrf-users Digest, Vol 138, Issue 16 - (no subject) (afwande juliet)
wrfhelp
wrfhelp at ucar.edu
Mon Feb 29 15:16:29 MST 2016
Afwande,
Under situation, I don't think you can obtain rainfall information. You
have to rerun the case.
WRFHELP
On 2/28/16 11:56 AM, afwande juliet wrote:
>
> Thanks , the bucket scheme was on when running experiment, and
> unfortunately, I didn't carry l_rainc and l_rainnc since I was running
> the model in another country and it seems the data was deleted from
> server to create space, how can I help in this case
>
> On Feb 27, 2016 2:40 AM, "wrfhelp" <wrfhelp at ucar.edu
> <mailto:wrfhelp at ucar.edu>> wrote:
>
>
> Afwande,
>
> I am not sure whether you have turned on the bucket scheme? If so,
> then you should calculate the total convective and
> resolvable-scale precipitation as below:
>
> RAINC=RAINC+I_RAINC*bucket_mm
> RAINNC=RAINNC+I_RAINNC*bucket_mm
>
> Daily and monthly total precipitation should be derived from the
> total precipitation. For example, if you have run WRF for two
> month, and you need to calculate monthly precipitation for the
> second month. Then you need to calculate accumulative
> precipitation at the end of the first month (suppose it is
> PRECIP1) and that at the end of second month (suppose it is
> PRECIP2). Monthly precipitation for the second month should be:
>
> PRECIP2 - PRECIP1
>
> WRFHELP
>
>
>
> On 2/26/16 10:50 AM, Jimy Dudhia wrote:
>> Maybe they have the bucket scheme on? If so, point them to the
>> user guide on the use of the bucket_mm variables like i_rainnc
>> Jimy
>>
>> ---------- Forwarded message ----------
>> From: *afwande juliet* <afwandej965 at gmail.com
>> <mailto:afwandej965 at gmail.com>>
>> Date: Fri, Feb 26, 2016 at 1:38 AM
>> Subject: Re: [Wrf-users] Wrf-users Digest, Vol 138, Issue 16 -
>> (no subject) (afwande juliet)
>> To: Felipe Neris <felipenc2 at gmail.com <mailto:felipenc2 at gmail.com>>
>> Cc: wrf users group <wrf-users at ucar.edu <mailto:wrf-users at ucar.edu>>
>>
>>
>> ok thanks
>> when I do investigation about my rainc&rainnc, i find that rainc
>> doesnt look cumulative (values decrease and increase randomly
>> within the time of simulation). Looking at rainc further, there
>> are big values upto 1000mm with some negative values in between.
>> However rainnc, looks cumulative and the values increase with
>> time and there are no negative values. Any other variables like
>> temp etc looks reasonable.
>>
>> And you know that precip=rainc+rainnc
>> What could be the problem and how can i correct this?
>>
>> On Fri, Feb 26, 2016 at 3:16 AM, Felipe Neris
>> <felipenc2 at gmail.com <mailto:felipenc2 at gmail.com>> wrote:
>>
>> Hi Juliet,
>> If I got your question right, I suppose the answer is: WRF
>> sums all precipitation that occurs since the beginning of the
>> simulation into the variables RAINNC (from microphysics
>> parameterization) and RAINC (from cumulus parameterization).
>> Therefor, if you want to have the accumulated precipitation
>> of a certain day, you must specify/set the corresponding time
>> for these variables and display.
>> Hope I could be of any help!
>>
>> Felipe Neris
>>
>> On Thu, Feb 25, 2016 at 3:45 PM, <wrf-users-request at ucar.edu
>> <mailto:wrf-users-request at ucar.edu>> wrote:
>>
>> Send Wrf-users mailing list submissions to
>> wrf-users at ucar.edu <mailto:wrf-users at ucar.edu>
>>
>> To subscribe or unsubscribe via the World Wide Web, visit
>> http://mailman.ucar.edu/mailman/listinfo/wrf-users
>> or, via email, send a message with subject or body 'help' to
>> wrf-users-request at ucar.edu
>> <mailto:wrf-users-request at ucar.edu>
>>
>> You can reach the person managing the list at
>> wrf-users-owner at ucar.edu <mailto:wrf-users-owner at ucar.edu>
>>
>> When replying, please edit your Subject line so it is
>> more specific
>> than "Re: Contents of Wrf-users digest..."
>>
>>
>> Today's Topics:
>>
>> 1. (no subject) (afwande juliet)
>> 2. Nesting and Domain Decomposition (Douglas Lowe)
>> 3. Re: Nesting and Domain Decomposition (Tabish Ansari)
>>
>>
>> ----------------------------------------------------------------------
>>
>> Message: 1
>> Date: Thu, 25 Feb 2016 15:45:08 +0300
>> From: afwande juliet <afwandej965 at gmail.com
>> <mailto:afwandej965 at gmail.com>>
>> Subject: [Wrf-users] (no subject)
>> To: wrf users group <wrf-users at ucar.edu
>> <mailto:wrf-users at ucar.edu>>, wrfhelp <wrfhelp at ucar.edu
>> <mailto:wrfhelp at ucar.edu>>
>> Message-ID:
>>
>> <CANVsOojVG3LzL1r82T0oKnBRAU-Nq2vVqHc5GKz4SSj9j0P3Cw at mail.gmail.com
>> <mailto:CANVsOojVG3LzL1r82T0oKnBRAU-Nq2vVqHc5GKz4SSj9j0P3Cw at mail.gmail.com>>
>> Content-Type: text/plain; charset="utf-8"
>>
>> I ask this again
>> I have WRF simulations for 1981. The model output is
>> 3hourly, i.e 8
>> timestep in a day
>> When I want daily values or monthly values, do i take
>> every 8th timestep to
>> be rain totals for each day and sum them up to get
>> monthly totals?
>> Do I have to divide the units *mm* by any number to get
>> mm/day ?
>>
>>
>>
>> thanks
>> -------------- next part --------------
>> An HTML attachment was scrubbed...
>> URL:
>> http://mailman.ucar.edu/pipermail/wrf-users/attachments/20160225/c10344ce/attachment-0001.html
>>
>> ------------------------------
>>
>> Message: 2
>> Date: Thu, 25 Feb 2016 13:59:29 +0000
>> From: Douglas Lowe <Douglas.Lowe at manchester.ac.uk
>> <mailto:Douglas.Lowe at manchester.ac.uk>>
>> Subject: [Wrf-users] Nesting and Domain Decomposition
>> To: "wrf-users at ucar.edu <mailto:wrf-users at ucar.edu>"
>> <wrf-users at ucar.edu <mailto:wrf-users at ucar.edu>>
>> Message-ID:
>>
>> <43E6B083008E774B87C4283E0FFA4E70012C7A0C91 at MBXP02.ds.man.ac.uk
>> <mailto:43E6B083008E774B87C4283E0FFA4E70012C7A0C91 at MBXP02.ds.man.ac.uk>>
>> Content-Type: text/plain; charset="us-ascii"
>>
>> Hi all,
>>
>> I'm running WRF-Chem with a nest of 3 domains, with the
>> settings listed below. I'd like to be
>> able to split this across as many processes as possible
>> in order to speed things up (currently
>> I'm only managing 3x real time, which isn't very good
>> when running multiday simulations).
>> Unfortunately I am finding that WRF hangs when calling
>> the photolysis driver for my 2nd domain
>> (which is the smallest of the domains) if I use too many
>> processors.
>>
>> The (relevant) model domain settings are:
>> max_dom = 3,
>> e_we = 134, 81, 91,
>> e_sn = 146, 81, 91,
>> e_vert = 41, 41, 41,
>> num_metgrid_levels = 38,
>> dx = 15000,3000,1000,
>> dy = 15000,3000,1000,
>>
>> WRF will run when I split over upto 168 processes (7
>> nodes on the ARCHER supercomputer),
>> but wont work if I split over 192 (or more) processes (8
>> nodes on ARCHER).
>>
>> Looking at the log messages I *think* that WRF is
>> splitting each domain into the same
>> number of patches, and sending one patch from each domain
>> to a single process for
>> analysis. However, this means that I am limited by the
>> smallest domain as to how many
>> patches I can split a domain into before we end up with
>> patches which are dwarved by
>> the halos around them.
>>
>> Would it not make more sense to be able to split each
>> domain into different numbers
>> of patches (so that each patch is of a similar size,
>> regardless of which domain it is from) and
>> send one patch from one domain to a single process (or,
>> perhaps, send more patches from the
>> outer domains to a single process, if needed for
>> balancing computational demands)? And
>> is there anyway for me to do this with WRF?
>>
>> Thanks,
>> Doug
>>
>> ------------------------------
>>
>> Message: 3
>> Date: Thu, 25 Feb 2016 16:37:26 +0000
>> From: Tabish Ansari <tabishumaransari at gmail.com
>> <mailto:tabishumaransari at gmail.com>>
>> Subject: Re: [Wrf-users] Nesting and Domain Decomposition
>> To: Douglas Lowe <Douglas.Lowe at manchester.ac.uk
>> <mailto:Douglas.Lowe at manchester.ac.uk>>
>> Cc: "wrf-users at ucar.edu <mailto:wrf-users at ucar.edu>"
>> <wrf-users at ucar.edu <mailto:wrf-users at ucar.edu>>
>> Message-ID:
>>
>> <CALLVTyvwh3nimJczxjfAy+gNML1PfJpDudJT8TQhkDQrCLnwZw at mail.gmail.com
>> <mailto:CALLVTyvwh3nimJczxjfAy%2BgNML1PfJpDudJT8TQhkDQrCLnwZw at mail.gmail.com>>
>> Content-Type: text/plain; charset="utf-8"
>>
>> Hi Doug,
>>
>> I'm not too knowledgeable in this but have some
>> literature which might be
>> of relevance. Please have a look at the attached files.
>>
>> Cheers,
>>
>> Tabish
>>
>> Tabish U Ansari
>> PhD student, Lancaster Environment Center
>> Lancaster Univeristy
>> Bailrigg, Lancaster,
>> LA1 4YW, United Kingdom
>>
>> On 25 February 2016 at 13:59, Douglas Lowe
>> <Douglas.Lowe at manchester.ac.uk
>> <mailto:Douglas.Lowe at manchester.ac.uk>>
>> wrote:
>>
>> > Hi all,
>> >
>> > I'm running WRF-Chem with a nest of 3 domains, with the
>> settings listed
>> > below. I'd like to be
>> > able to split this across as many processes as possible
>> in order to speed
>> > things up (currently
>> > I'm only managing 3x real time, which isn't very good
>> when running
>> > multiday simulations).
>> > Unfortunately I am finding that WRF hangs when calling
>> the photolysis
>> > driver for my 2nd domain
>> > (which is the smallest of the domains) if I use too
>> many processors.
>> >
>> > The (relevant) model domain settings are:
>> > max_dom = 3,
>> > e_we = 134, 81, 91,
>> > e_sn = 146, 81, 91,
>> > e_vert = 41, 41, 41,
>> > num_metgrid_levels = 38,
>> > dx = 15000,3000,1000,
>> > dy = 15000,3000,1000,
>> >
>> > WRF will run when I split over upto 168 processes (7
>> nodes on the ARCHER
>> > supercomputer),
>> > but wont work if I split over 192 (or more) processes
>> (8 nodes on ARCHER).
>> >
>> > Looking at the log messages I *think* that WRF is
>> splitting each domain
>> > into the same
>> > number of patches, and sending one patch from each
>> domain to a single
>> > process for
>> > analysis. However, this means that I am limited by the
>> smallest domain as
>> > to how many
>> > patches I can split a domain into before we end up with
>> patches which are
>> > dwarved by
>> > the halos around them.
>> >
>> > Would it not make more sense to be able to split each
>> domain into
>> > different numbers
>> > of patches (so that each patch is of a similar size,
>> regardless of which
>> > domain it is from) and
>> > send one patch from one domain to a single process (or,
>> perhaps, send more
>> > patches from the
>> > outer domains to a single process, if needed for
>> balancing computational
>> > demands)? And
>> > is there anyway for me to do this with WRF?
>> >
>> > Thanks,
>> > Doug
>> > _______________________________________________
>> > Wrf-users mailing list
>> > Wrf-users at ucar.edu <mailto:Wrf-users at ucar.edu>
>> > http://mailman.ucar.edu/mailman/listinfo/wrf-users
>> >
>> -------------- next part --------------
>> An HTML attachment was scrubbed...
>> URL:
>> http://mailman.ucar.edu/pipermail/wrf-users/attachments/20160225/02afb9aa/attachment.html
>> -------------- next part --------------
>> A non-text attachment was scrubbed...
>> Name: WRF-HPC.pdf
>> Type: application/pdf
>> Size: 243897 bytes
>> Desc: not available
>> Url :
>> http://mailman.ucar.edu/pipermail/wrf-users/attachments/20160225/02afb9aa/attachment.pdf
>> -------------- next part --------------
>> A non-text attachment was scrubbed...
>> Name: WRF-chapter-multicore.pdf
>> Type: application/pdf
>> Size: 230093 bytes
>> Desc: not available
>> Url :
>> http://mailman.ucar.edu/pipermail/wrf-users/attachments/20160225/02afb9aa/attachment-0001.pdf
>> -------------- next part --------------
>> A non-text attachment was scrubbed...
>> Name: CUDA-WRF_ppt.pdf
>> Type: application/pdf
>> Size: 2314206 bytes
>> Desc: not available
>> Url :
>> http://mailman.ucar.edu/pipermail/wrf-users/attachments/20160225/02afb9aa/attachment-0002.pdf
>>
>> ------------------------------
>>
>> _______________________________________________
>> Wrf-users mailing list
>> Wrf-users at ucar.edu <mailto:Wrf-users at ucar.edu>
>> http://mailman.ucar.edu/mailman/listinfo/wrf-users
>>
>>
>> End of Wrf-users Digest, Vol 138, Issue 16
>> ******************************************
>>
>>
>>
>> _______________________________________________
>> Wrf-users mailing list
>> Wrf-users at ucar.edu <mailto:Wrf-users at ucar.edu>
>> http://mailman.ucar.edu/mailman/listinfo/wrf-users
>>
>>
>>
>> _______________________________________________
>> Wrf-users mailing list
>> Wrf-users at ucar.edu <mailto:Wrf-users at ucar.edu>
>> http://mailman.ucar.edu/mailman/listinfo/wrf-users
>>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: http://mailman.ucar.edu/pipermail/wrf-users/attachments/20160229/6ec5f2c8/attachment-0001.html
More information about the Wrf-users
mailing list