diff options
author | 2019-10-23 15:07:01 -0700 | |
---|---|---|
committer | 2019-10-23 15:07:01 -0700 | |
commit | 4e1de98954f7f0acef04efe092a43256eabb8a91 (patch) | |
tree | c07a17e6375ff2480e135c534a6da511a0d7b094 /Tools/performance_tests/run_automated.py | |
parent | cb015e5aae22d6688583b8cd23ab72df10ac4b12 (diff) | |
parent | 51cf0b010b5d556cf7f415eec0938a313ca181f4 (diff) | |
download | WarpX-4e1de98954f7f0acef04efe092a43256eabb8a91.tar.gz WarpX-4e1de98954f7f0acef04efe092a43256eabb8a91.tar.zst WarpX-4e1de98954f7f0acef04efe092a43256eabb8a91.zip |
Merge pull request #484 from MaxThevenet/perftest2
Perftest2
Diffstat (limited to 'Tools/performance_tests/run_automated.py')
-rw-r--r-- | Tools/performance_tests/run_automated.py | 16 |
1 files changed, 7 insertions, 9 deletions
diff --git a/Tools/performance_tests/run_automated.py b/Tools/performance_tests/run_automated.py index 1a4184013..bec3f28d8 100644 --- a/Tools/performance_tests/run_automated.py +++ b/Tools/performance_tests/run_automated.py @@ -97,7 +97,7 @@ if args.automated == True: pull_3_repos = True recompile = True if machine == 'summit': - compiler = 'pgi' + compiler = 'gnu' architecture = 'gpu' # List of tests to perform @@ -121,12 +121,13 @@ perf_logs_repo = source_dir_base + 'perf_logs/' compiler_name = {'intel': 'intel', 'gnu': 'gcc', 'pgi':'pgi'} module_Cname = {'cpu': 'haswell', 'knl': 'knl,quad,cache', 'gpu':''} csv_file = {'cori':'cori_knl.csv', 'summit':'summit.csv'} -cwd = os.getcwd() + '/' +# cwd = os.getcwd() + '/' +cwd = warpx_dir + 'Tools/performance_tests/' +print('cwd = ' + cwd) bin_dir = cwd + 'Bin/' bin_name = executable_name(compiler, architecture) log_dir = cwd -perf_database_file = cwd + perf_database_file day = time.strftime('%d') month = time.strftime('%m') year = time.strftime('%Y') @@ -247,15 +248,14 @@ for n_node in n_node_list: df_newline['inputs_content'] = get_file_content( filename=cwd+current_run.input_file ) # Load file perf_database_file if exists, and # append with results from this scan - if os.path.exists(perf_database_file): - # df_base = pd.read_hdf(perf_database_file, 'all_data', format='table') - df_base = pd.read_hdf(perf_database_file, 'all_data') + if os.path.exists(perf_logs_repo + '/logs_hdf5/' + perf_database_file): + df_base = pd.read_hdf(perf_logs_repo + '/logs_hdf5/' + perf_database_file, 'all_data') updated_df = df_base.append(df_newline, ignore_index=True) else: updated_df = df_newline # Write dataframe to file perf_database_file # (overwrite if file exists) - updated_df.to_hdf(perf_database_file, key='all_data', mode='w') + updated_df.to_hdf(perf_logs_repo + '/logs_hdf5/' + perf_database_file, key='all_data', mode='w', format='table') # Extract sub-set of pandas data frame, write it to # csv file and copy this file to perf_logs repo @@ -266,8 +266,6 @@ if update_perf_log_repo: if push_on_perf_log_repo: git_repo.git.stash('save') git_repo.git.pull() - # move csv file to perf_logs repon and commit the new version - shutil.move( perf_database_file, perf_logs_repo + '/logs_hdf5/' + perf_database_file ) os.chdir( perf_logs_repo ) sys.path.append('./') import generate_index_html |