Mercurial > repos > tduigou > data_manager_cache
diff data_manager/cache_fetcher.py @ 1:35c33747b9e3 draft
"planemo upload commit f40274f6b9f6a15eb4022aab21286d4c96cd8475-dirty"
| author | tduigou |
|---|---|
| date | Tue, 05 Jul 2022 12:57:38 +0000 |
| parents | e0b92d203870 |
| children |
line wrap: on
line diff
--- a/data_manager/cache_fetcher.py Mon Jul 04 13:28:30 2022 +0000 +++ b/data_manager/cache_fetcher.py Tue Jul 05 12:57:38 2022 +0000 @@ -10,6 +10,7 @@ ) import argparse +import json def download( url: str, @@ -36,18 +37,43 @@ f.close() return file +def download_entries(url, filename, workdir): + full_filename = os_path.abspath(os_path.join(workdir, filename+".json.gz")) + + download(url+filename +".json.gz", full_filename) + data_manager_entry = {} + data_manager_entry["value"] = filename + data_manager_entry["name"] = filename + data_manager_entry["path"] = full_filename + + # Make sure that less than 10 requests per second, as required by host (http://bigg.ucsd.edu/data_access) + #if ix % 5 == 0: + # time.sleep(1) + yield data_manager_entry + + parser = argparse.ArgumentParser(description="Download a cache file") -parser.add_argument('-u','--url', required=True, default=None, type=str, help="URL the file is downloaded from") +parser.add_argument('-f','--filename', required=True, default=None, type=str, help="Cache filename to download") parser.add_argument('-o','--outfile', required=True, default=None, type=str, help="A filename where the downloaded file has stored into") args = parser.parse_args() -url= args.url #"https://gitlab.com/breakthewall/rrCache-data/-/raw/master/" -filename= os_path.basename(args.outfile) #"cid_strc.json.gz" -cache_dir=os_path.dirname(args.outfile) #'${GALAXY_DATA_MANAGER_DATA_PATH}'+'/rpextractsink/cache/' -full_filename=os_path.join(cache_dir,filename) #'${GALAXY_DATA_MANAGER_DATA_PATH}'+'/rpextractsink/cache/cid_strc.json.gz' +url= "https://gitlab.com/breakthewall/rrCache-data/-/raw/master/" +filename= args.filename + + +data_manager_json = {"data_tables": {}} +with open(args.outfile) as fh: + params = json.load(fh) -if not os_path.isdir(cache_dir): - os_mkdir(cache_dir) +workdir = params["output_data"][0]["extra_files_path"] +os_mkdir(workdir) + +#full_filename=os_path.join(workdir,filename) #'${GALAXY_DATA_MANAGER_DATA_PATH}'+'/rpextractsink/cache/cid_strc.json.gz' + -download(url+filename, full_filename) \ No newline at end of file +entries = list(download_entries(url, filename, workdir)) + +data_manager_json["data_tables"]["cache"] = entries +with open(args.outfile, "w") as fh: + json.dump(data_manager_json, fh, sort_keys=True) \ No newline at end of file
