diff data_manager/cache_fetcher.py @ 1:35c33747b9e3 draft

"planemo upload commit f40274f6b9f6a15eb4022aab21286d4c96cd8475-dirty"
author tduigou
date Tue, 05 Jul 2022 12:57:38 +0000
parents e0b92d203870
children
line wrap: on
line diff
--- a/data_manager/cache_fetcher.py	Mon Jul 04 13:28:30 2022 +0000
+++ b/data_manager/cache_fetcher.py	Tue Jul 05 12:57:38 2022 +0000
@@ -10,6 +10,7 @@
 )
 
 import argparse
+import json
 
 def download(
     url: str,
@@ -36,18 +37,43 @@
     f.close()
     return file
 
+def download_entries(url, filename, workdir):
+    full_filename = os_path.abspath(os_path.join(workdir, filename+".json.gz"))
+
+    download(url+filename +".json.gz", full_filename)
+    data_manager_entry = {}
+    data_manager_entry["value"] = filename
+    data_manager_entry["name"] = filename
+    data_manager_entry["path"] = full_filename
+
+    # Make sure that less than 10 requests per second, as required by host (http://bigg.ucsd.edu/data_access)
+    #if ix % 5 == 0:
+    #    time.sleep(1)
+    yield data_manager_entry
+
+
 parser = argparse.ArgumentParser(description="Download a cache file")
-parser.add_argument('-u','--url', required=True, default=None, type=str, help="URL the file is downloaded from")
+parser.add_argument('-f','--filename', required=True, default=None, type=str, help="Cache filename to download")
 parser.add_argument('-o','--outfile', required=True, default=None, type=str, help="A filename where the downloaded file has stored into")
 
 args = parser.parse_args()
 
-url= args.url #"https://gitlab.com/breakthewall/rrCache-data/-/raw/master/"
-filename= os_path.basename(args.outfile) #"cid_strc.json.gz"
-cache_dir=os_path.dirname(args.outfile) #'${GALAXY_DATA_MANAGER_DATA_PATH}'+'/rpextractsink/cache/'
-full_filename=os_path.join(cache_dir,filename) #'${GALAXY_DATA_MANAGER_DATA_PATH}'+'/rpextractsink/cache/cid_strc.json.gz'
+url= "https://gitlab.com/breakthewall/rrCache-data/-/raw/master/"
+filename= args.filename
+
+
+data_manager_json = {"data_tables": {}}
+with open(args.outfile) as fh:
+    params = json.load(fh)
 
-if not os_path.isdir(cache_dir):
-    os_mkdir(cache_dir)
+workdir = params["output_data"][0]["extra_files_path"]
+os_mkdir(workdir)
+
+#full_filename=os_path.join(workdir,filename) #'${GALAXY_DATA_MANAGER_DATA_PATH}'+'/rpextractsink/cache/cid_strc.json.gz'
+
 
-download(url+filename, full_filename)
\ No newline at end of file
+entries = list(download_entries(url, filename, workdir))
+
+data_manager_json["data_tables"]["cache"] = entries
+with open(args.outfile, "w") as fh:
+    json.dump(data_manager_json, fh, sort_keys=True)
\ No newline at end of file