[Wrf-users] memory space for long simulations
Don Morton
don.morton at borealscicomp.com
Thu Nov 12 09:10:40 MST 2015
Hello,
I'm guessing that you could compile with WRFIO_NCD_LARGE_FILE_SUPPORT and
get around the restriction. Please see Q17 in the following:
http://www2.mmm.ucar.edu/wrf/users/FAQ_files/FAQ_wrf_runtime.html
You might also need to make sure your NetCDF is compiled for large memory
support (it might be by default, but I can't remember for sure).
---
Don Morton, Owner/Manager
Boreal Scientific Computing LLC
Fairbanks, Alaska USA
http://www.borealscicomp.com/
http://www.borealscicomp.com/Miscellaneous/MortonBio/
On Thu, Nov 12, 2015 at 10:55 AM, afwande juliet <afwandej965 at gmail.com>
wrote:
> Hi all users
>
> How do users go about this issue of frames_per_outfile? I am doing 5years
> simulations for 3 different cases. The interval_seconds =21600(6hours).
> That means each case there will be 7300files*3=22000files(total files).
> I read somewhere that max space for frames_per_outputfile=2GB. I guess if
> the 5years run is complete the total disk space will be 20T which the
> cluster am using has .
> I don't want to have many files; If possible all files in one big file BUT
> I am afraid doing this will make model stop giving output due to
> insufficient memory.
>
> Can someone help me how to go about this!!
>
>
> &time_control
> !run_days = 0,
> !run_hours = 96,
> !run_minutes = 0,
> !run_seconds = 0,
> start_year = 1997,
> start_month = 02,
> start_day = 01,
> start_hour = 00,
> start_minute = 00,
> start_second = 00,
> end_year = 2006,
> end_month = 12,
> end_day = 31,
> end_hour = 00,
> end_minute = 00,
> end_second = 00
> interval_seconds = 21600
> input_from_file = .false.,
> history_interval = 180,
> frames_per_outfile = 100,
> restart = .false.,
> restart_interval = 1440,
> io_form_history = 2
> io_form_restart = 2
> io_form_input = 2
> io_form_boundary = 2
> debug_level = 100
> auxinput4_inname = "wrflowinp_d<domain>"
> auxinput4_interval = 360,
> io_form_auxinput4 = 2,
> /
> &domains
> time_step = 200,
> time_step_fract_num = 0,
> time_step_fract_den = 1,
> max_dom = 1,
> e_we = 250,
> e_sn = 150,
> e_vert = 51,
> p_top_requested = 5000,
> num_metgrid_levels = 38,
> num_metgrid_soil_levels = 4,
> dx = 36000, 12000, 3333.33,
> dy = 36000, 12000, 3333.33,
> grid_id = 1,
> parent_id = 0,
> i_parent_start = 1,
> j_parent_start = 1,
> parent_grid_ratio = 1,
> parent_time_step_ratio = 1,
> feedback = 1,
> smooth_option = 0,
> /
> &physics
> mp_physics = 6,
> ra_lw_physics = 4,
> ra_sw_physics = 4,
> radt = 30,
> sf_sfclay_physics = 1,
> sf_surface_physics = 2,
> bl_pbl_physics = 1,
> bldt = 0,
> cu_physics = 1,
> cudt = 5,
> kfeta_trigger = 1,
> cu_rad_feedback = .true.,
> isfflx = 1,
> sst_update = 1,
> tmn_update = 1,
> sst_skin = 1,
> bucket_mm = 100.0,
> bucket_J = 1.e9,
> ifsnow = 1,
> icloud = 1,
> surface_input_source = 1,
> num_soil_layers = 4,
> sf_urban_physics = 0,
> /
> &fdda
> grid_fdda = 2,
> gfdda_inname = "wrffdda_d<domain>",
> gfdda_interval_m = 360,
> gfdda_end_h = 52560,
> io_form_gfdda = 2,
> fgdt = 0,
> if_no_pbl_nudging_uv = 1,
> if_no_pbl_nudging_t = 1,
> if_no_pbl_nudging_q = 1,
> if_no_pbl_nudging_ph = 1,
> if_zfac_uv = 1,
> k_zfac_uv = 18,
> if_zfac_t = 1,
> k_zfac_t = 18,
> if_zfac_q = 1,
> k_zfac_q = 18,
> if_zfac_ph = 1,
> k_zfac_ph = 18,
> guv = 0.0003,
> gt = 0.0003,
> gq = 0.0003,
> gph = 0.0003,
> xwavenum = 9,
> ywavenum = 5,
> if_ramping = 0,
> dtramp_min = 60.0,
> /
> &dynamics
> w_damping = 1,
> diff_opt = 1, 1, 1,
> km_opt = 4, 4, 4,
> diff_6th_opt = 0, 0, 0,
> diff_6th_factor = 0.12, 0.12, 0.12,
> base_temp = 290.
> damp_opt = 3,
> zdamp = 5000., 5000., 5000.,
> dampcoef = 0.2, 0.2, 0.2
> khdif = 0, 0, 0,
> kvdif = 0, 0, 0,
> non_hydrostatic = .true., .true., .true.,
> moist_adv_opt = 1, 1, 1,
> scalar_adv_opt = 1, 1, 1,
> /
> &bdy_control
> spec_bdy_width = 10,
> spec_zone = 1,
> relax_zone = 9,
> specified = .true., .false.,.false.,
> nested = .false., .true., .true.,
> spec_exp = 0.33,
> /
> &grib2
> /
> &namelist_quilt
> nio_tasks_per_group = 0,
> nio_groups = 1,
> /
>
>
>
> Thanks afwande
>
> 1
>
>
> _______________________________________________
> Wrf-users mailing list
> Wrf-users at ucar.edu
> http://mailman.ucar.edu/mailman/listinfo/wrf-users
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: http://mailman.ucar.edu/pipermail/wrf-users/attachments/20151112/309ea3ed/attachment.html
More information about the Wrf-users
mailing list