Mercurial > repos > fabio > gdcwebapp
changeset 32:1edc869cd008 draft
Uploaded 20170607
author | fabio |
---|---|
date | Wed, 07 Jun 2017 18:01:57 -0400 |
parents | 9cb5e4f12ce5 |
children | 228038cd0683 |
files | ._gdcwebapp.xml ._json_collect_data_source.py gdcwebapp.xml json_collect_data_source.py json_data_source_mod.py |
diffstat | 5 files changed, 283 insertions(+), 282 deletions(-) [+] |
line wrap: on
line diff
--- a/gdcwebapp.xml Tue Jun 06 23:04:41 2017 -0400 +++ b/gdcwebapp.xml Wed Jun 07 18:01:57 2017 -0400 @@ -11,7 +11,7 @@ <command> <![CDATA[ mkdir -p tmp && - python ${__tool_directory__}/json_data_source_mod.py "${__app__.config.output_size_limit}" --json_param_file "${output1}" --path "." --appdata "tmp" + python ${__tool_directory__}/json_collect_data_source.py "${__app__.config.output_size_limit}" --json_param_file "${output1}" --path "." --appdata "tmp" ]]> </command> <inputs check_values="False" action="http://bioinf.iasi.cnr.it/gdcwebapp/app.php">
--- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/json_collect_data_source.py Wed Jun 07 18:01:57 2017 -0400 @@ -0,0 +1,282 @@ +#!/usr/bin/env python +import json +import optparse +import urllib +import os.path +import os +from operator import itemgetter +import tarfile + +__version__ = "1.0.0" +CHUNK_SIZE = 2**20 #1mb +VALID_CHARS = '.-()[]0123456789abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ ' + + +def splitext(path): + for ext in ['.tar.gz', '.tar.bz2']: + if path.endswith(ext): + path, ext = path[:-len(ext)], path[-len(ext):] + break + else: + path, ext = os.path.splitext(path) + return path, ext[1:] + + +def chunk_write( source_stream, target_stream, source_method = "read", target_method="write" ): + source_method = getattr( source_stream, source_method ) + target_method = getattr( target_stream, target_method ) + while True: + chunk = source_method( CHUNK_SIZE ) + if chunk: + target_method( chunk ) + else: + break + + +def deconstruct_multi_filename( multi_filename ): + keys = [ 'primary', 'id', 'name', 'visible', 'file_type' ] + return ( dict( zip( keys, multi_filename.split('_') ) ) ) + + +def construct_multi_filename( id, name, file_type ): + """ Implementation of *Number of Output datasets cannot be determined until tool run* from documentation_. + .. _documentation: http://wiki.galaxyproject.org/Admin/Tools/Multiple%20Output%20Files + """ + filename = "%s_%s_%s_%s_%s" % ( 'primary', id, name, 'visible', file_type ) + return filename + + +def download_from_query( query_data, target_output_filename ): + """ Download file from the json data and write it to target_output_filename. + """ + query_url = query_data.get( 'url' ) + query_file_type = query_data.get( 'extension' ) + query_stream = urllib.urlopen( query_url ) + output_stream = open( target_output_filename, 'wb' ) + chunk_write( query_stream, output_stream ) + query_stream.close() + output_stream.close() + +def store_file_from_archive( file_object, target_output_filename, isString=False ): + """ Store file after extracting from archive and organize them as a collection using the structure + (collection-name)_(file-name).ext as file name + """ + output_stream = open( target_output_filename, 'wb' ) + #chunk_write( file_object.read(), output_stream ) + if not isString: + output_stream.write(file_object.read()) + else: + output_stream.write(file_object) + output_stream.close() + + +def download_extra_data( query_ext_data, base_path ): + """ Download any extra data defined in the JSON. + NOTE: the "path" value is a relative path to the file on our + file system. This is slightly dangerous and we should make every effort + to avoid a malicious absolute path to write the file elsewhere on the + filesystem. + """ + for ext_data in query_ext_data: + if not os.path.exists( base_path ): + os.mkdir( base_path ) + query_stream = urllib.urlopen( ext_data.get( 'url' ) ) + ext_path = ext_data.get( 'path' ) + os.makedirs( os.path.normpath( '/'.join( [ base_path, os.path.dirname( ext_path ) ] ) ) ) + output_stream = open( os.path.normpath( '/'.join( [ base_path, ext_path ] ) ), 'wb' ) + chunk_write( query_stream, output_stream ) + query_stream.close() + output_stream.close() + + +def metadata_to_json( dataset_id, metadata, filename, ds_type='dataset', primary=False): + """ Return line separated JSON """ + meta_dict = dict( type = ds_type, + ext = metadata.get( 'extension' ), + filename = filename, + name = metadata.get( 'name' ), + metadata = metadata.get( 'metadata', {} ) ) + if metadata.get( 'extra_data', None ): + meta_dict[ 'extra_files' ] = '_'.join( [ filename, 'files' ] ) + if primary: + meta_dict[ 'base_dataset_id' ] = dataset_id + else: + meta_dict[ 'dataset_id' ] = dataset_id + return "%s\n" % json.dumps( meta_dict ) + + +def walk_on_archive(target_output_filename, check_ext, archive_name, appdata_path, db_key="?"): + archive_name = archive_name.replace("_", "-").replace(".", "-") + with tarfile.open( target_output_filename, check_ext ) as tf: + for entry in tf: + if entry.isfile(): + fileobj = tf.extractfile( entry ) + # reserve the underscore for the collection searator + filename = os.path.basename( entry.name ).replace("_", "-") + extension = splitext( filename )[1] + # pattern: (?P<identifier_0>[^_]+)_(?P<identifier_1>[^_]+)_(?P<ext>[^_]+)_(?P<dbkey>[^_]+) + if (len(extension) > 0): + filename = (filename[0:len(filename)-(len(extension)+1)]).replace(".", "-") + "." + extension + "_" + extension + else: + extension = "auto" + filename_with_collection_prefix = archive_name + "_" + filename + "_" + db_key + target_entry_output_filename = os.path.join(appdata_path, filename_with_collection_prefix) + store_file_from_archive( fileobj, target_entry_output_filename ) + return True + + +def download_files_and_write_metadata(query_item, json_params, output_base_path, metadata_parameter_file, primary, appdata_path, options, args): + """ Main work function that operates on the JSON representation of + one dataset and its metadata. Returns True. + """ + dataset_url, output_filename, \ + extra_files_path, file_name, \ + ext, out_data_name, \ + hda_id, dataset_id = set_up_config_values(json_params) + extension = query_item.get( 'extension' ) + url = query_item.get( 'url' ) + filename = query_item.get( 'name' ) + + check_ext = "" + if ( url.endswith( "gz" ) ): + check_ext = "r:gz" + elif ( url.endswith( "bz2" ) ): + check_ext = "r:bz2" + elif ( url.endswith( "tar" ) ): + check_ext = "r:" + isArchive = bool( check_ext and check_ext.strip() ) + + extra_data = query_item.get( 'extra_data', None ) + if primary: + filename = ''.join( c in VALID_CHARS and c or '-' for c in filename ) + name = construct_multi_filename( hda_id, filename, extension ) + target_output_filename = os.path.normpath( '/'.join( [ output_base_path, name ] ) ) + if isArchive is False: + metadata_parameter_file.write( metadata_to_json( dataset_id, query_item, + target_output_filename, + ds_type='new_primary_dataset', + primary=primary) ) + else: + target_output_filename = output_filename + if isArchive is False: + metadata_parameter_file.write( metadata_to_json( dataset_id, query_item, + target_output_filename, + ds_type='dataset', + primary=primary) ) + + if isArchive is False: + download_from_query( query_item, target_output_filename ) + else: + target_output_path = os.path.join(appdata_path, filename) + download_from_query( query_item, target_output_path ) + if extra_data: + extra_files_path = ''.join( [ target_output_filename, 'files' ] ) + download_extra_data( extra_data, extra_files_path ) + + """ the following code handles archives and decompress them in a collection """ + if ( isArchive ): + db_key = "?" + archive_metadata = query_item.get( 'metadata', None ) + if archive_metadata is not None: + try: + db_key = archive_metadata.get( 'db_key' ) + except: + pass + walk_on_archive(target_output_path, check_ext, filename, appdata_path, db_key) + + return True + + +def set_up_config_values(json_params): + """ Parse json_params file and return a tuple of necessary configuration + values. + """ + datasource_params = json_params.get( 'param_dict' ) + dataset_url = datasource_params.get( 'URL' ) + output_filename = datasource_params.get( 'output1', None ) + output_data = json_params.get( 'output_data' ) + extra_files_path, file_name, ext, out_data_name, hda_id, dataset_id = \ + itemgetter('extra_files_path', 'file_name', 'ext', 'out_data_name', 'hda_id', 'dataset_id')(output_data[0]) + return (dataset_url, output_filename, + extra_files_path, file_name, + ext, out_data_name, + hda_id, dataset_id) + + +def download_from_json_data( options, args ): + """ Parse the returned JSON data and download files. Write metadata + to flat JSON file. + """ + output_base_path = options.path + appdata_path = options.appdata + if not os.path.exists(appdata_path): + os.makedirs(appdata_path) + + # read tool job configuration file and parse parameters we need + json_params = json.loads( open( options.json_param_file, 'r' ).read() ) + print("json_params: "+str(json_params)) + + dataset_url, output_filename, \ + extra_files_path, file_name, \ + ext, out_data_name, \ + hda_id, dataset_id = set_up_config_values(json_params) + # line separated JSON file to contain all dataset metadata + metadata_parameter_file = open( json_params['job_config']['TOOL_PROVIDED_JOB_METADATA_FILE'], 'wb' ) + + # get JSON response from data source + # TODO: make sure response is not enormous + query_params = json.loads(urllib.urlopen( dataset_url ).read()) + # download and write files + primary = False + #primary = True + # query_item, hda_id, output_base_path, dataset_id + for query_item in query_params: + if isinstance( query_item, list ): + # TODO: do something with the nested list as a collection + for query_subitem in query_item: + primary = download_files_and_write_metadata(query_subitem, json_params, output_base_path, + metadata_parameter_file, primary, appdata_path, options, args) + + elif isinstance( query_item, dict ): + primary = download_files_and_write_metadata(query_item, json_params, output_base_path, + metadata_parameter_file, primary, appdata_path, options, args) + metadata_parameter_file.close() + +def __main__(): + """ Read the JSON return from a data source. Parse each line and request + the data, download to "newfilepath", and write metadata. + + Schema + ------ + + [ {"url":"http://url_of_file", + "name":"encode WigData", + "extension":"wig", + "metadata":{"db_key":"hg19"}, + "extra_data":[ {"url":"http://url_of_ext_file", + "path":"rel/path/to/ext_file"} + ] + } + ] + + """ + # Parse the command line options + usage = "Usage: json_data_source_mod.py max_size --json_param_file filename [options]" + parser = optparse.OptionParser(usage = usage) + parser.add_option("-j", "--json_param_file", type="string", + action="store", dest="json_param_file", help="json schema return data") + parser.add_option("-p", "--path", type="string", + action="store", dest="path", help="new file path") + parser.add_option("-a", "--appdata", type="string", + action="store", dest="appdata", help="appdata folder name") + parser.add_option("-v", "--version", action="store_true", dest="version", + default=False, help="display version and exit") + + (options, args) = parser.parse_args() + if options.version: + print __version__ + else: + download_from_json_data( options, args ) + + +if __name__ == "__main__": __main__()
--- a/json_data_source_mod.py Tue Jun 06 23:04:41 2017 -0400 +++ /dev/null Thu Jan 01 00:00:00 1970 +0000 @@ -1,281 +0,0 @@ -#!/usr/bin/env python -import json -import optparse -import urllib -import os.path -import os -from operator import itemgetter -import tarfile - -__version__ = "1.0.0" -CHUNK_SIZE = 2**20 #1mb -VALID_CHARS = '.-()[]0123456789abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ ' - - -def splitext(path): - for ext in ['.tar.gz', '.tar.bz2']: - if path.endswith(ext): - path, ext = path[:-len(ext)], path[-len(ext):] - break - else: - path, ext = os.path.splitext(path) - return path, ext[1:] - - -def chunk_write( source_stream, target_stream, source_method = "read", target_method="write" ): - source_method = getattr( source_stream, source_method ) - target_method = getattr( target_stream, target_method ) - while True: - chunk = source_method( CHUNK_SIZE ) - if chunk: - target_method( chunk ) - else: - break - - -def deconstruct_multi_filename( multi_filename ): - keys = [ 'primary', 'id', 'name', 'visible', 'file_type' ] - return ( dict( zip( keys, multi_filename.split('_') ) ) ) - - -def construct_multi_filename( id, name, file_type ): - """ Implementation of *Number of Output datasets cannot be determined until tool run* from documentation_. - .. _documentation: http://wiki.galaxyproject.org/Admin/Tools/Multiple%20Output%20Files - """ - filename = "%s_%s_%s_%s_%s" % ( 'primary', id, name, 'visible', file_type ) - return filename - - -def download_from_query( query_data, target_output_filename ): - """ Download file from the json data and write it to target_output_filename. - """ - query_url = query_data.get( 'url' ) - query_file_type = query_data.get( 'extension' ) - query_stream = urllib.urlopen( query_url ) - output_stream = open( target_output_filename, 'wb' ) - chunk_write( query_stream, output_stream ) - query_stream.close() - output_stream.close() - -def store_file_from_archive( file_object, target_output_filename, isString=False ): - """ Store file after extracting from archive and organize them as a collection using the structure - (collection-name)_(file-name).ext as file name - """ - output_stream = open( target_output_filename, 'wb' ) - #chunk_write( file_object.read(), output_stream ) - if not isString: - output_stream.write(file_object.read()) - else: - output_stream.write(file_object) - output_stream.close() - - -def download_extra_data( query_ext_data, base_path ): - """ Download any extra data defined in the JSON. - NOTE: the "path" value is a relative path to the file on our - file system. This is slightly dangerous and we should make every effort - to avoid a malicious absolute path to write the file elsewhere on the - filesystem. - """ - for ext_data in query_ext_data: - if not os.path.exists( base_path ): - os.mkdir( base_path ) - query_stream = urllib.urlopen( ext_data.get( 'url' ) ) - ext_path = ext_data.get( 'path' ) - os.makedirs( os.path.normpath( '/'.join( [ base_path, os.path.dirname( ext_path ) ] ) ) ) - output_stream = open( os.path.normpath( '/'.join( [ base_path, ext_path ] ) ), 'wb' ) - chunk_write( query_stream, output_stream ) - query_stream.close() - output_stream.close() - - -def metadata_to_json( dataset_id, metadata, filename, ds_type='dataset', primary=False): - """ Return line separated JSON """ - meta_dict = dict( type = ds_type, - ext = metadata.get( 'extension' ), - filename = filename, - name = metadata.get( 'name' ), - metadata = metadata.get( 'metadata', {} ) ) - if metadata.get( 'extra_data', None ): - meta_dict[ 'extra_files' ] = '_'.join( [ filename, 'files' ] ) - if primary: - meta_dict[ 'base_dataset_id' ] = dataset_id - else: - meta_dict[ 'dataset_id' ] = dataset_id - return "%s\n" % json.dumps( meta_dict ) - - -def walk_on_archive(target_output_filename, check_ext, archive_name, appdata_path, db_key="?"): - archive_name = archive_name.replace("_", "-").replace(".", "-") - with tarfile.open( target_output_filename, check_ext ) as tf: - for entry in tf: - if entry.isfile(): - fileobj = tf.extractfile( entry ) - # reserve the underscore for the collection searator - filename = os.path.basename( entry.name ).replace("_", "-") - extension = splitext( filename )[1] - # pattern: (?P<identifier_0>[^_]+)_(?P<identifier_1>[^_]+)_(?P<ext>[^_]+)_(?P<dbkey>[^_]+) - if (len(extension) > 0): - filename = (filename[0:len(filename)-(len(extension)+1)]).replace(".", "-") + "." + extension + "_" + extension - else: - extension = "auto" - filename_with_collection_prefix = archive_name + "_" + filename + "_" + db_key - target_entry_output_filename = os.path.join(appdata_path, filename_with_collection_prefix) - store_file_from_archive( fileobj, target_entry_output_filename ) - return True - - -def download_files_and_write_metadata(query_item, json_params, output_base_path, metadata_parameter_file, primary, appdata_path, options, args): - """ Main work function that operates on the JSON representation of - one dataset and its metadata. Returns True. - """ - dataset_url, output_filename, \ - extra_files_path, file_name, \ - ext, out_data_name, \ - hda_id, dataset_id = set_up_config_values(json_params) - extension = query_item.get( 'extension' ) - url = query_item.get( 'url' ) - filename = query_item.get( 'name' ) - - check_ext = "" - if ( url.endswith( "gz" ) ): - check_ext = "r:gz" - elif ( url.endswith( "bz2" ) ): - check_ext = "r:bz2" - elif ( url.endswith( "tar" ) ): - check_ext = "r:" - isArchive = bool( check_ext and check_ext.strip() ) - - extra_data = query_item.get( 'extra_data', None ) - if primary: - filename = ''.join( c in VALID_CHARS and c or '-' for c in filename ) - name = construct_multi_filename( hda_id, filename, extension ) - target_output_filename = os.path.normpath( '/'.join( [ output_base_path, name ] ) ) - if isArchive is False: - metadata_parameter_file.write( metadata_to_json( dataset_id, query_item, - target_output_filename, - ds_type='new_primary_dataset', - primary=primary) ) - else: - target_output_filename = output_filename - if isArchive is False: - metadata_parameter_file.write( metadata_to_json( dataset_id, query_item, - target_output_filename, - ds_type='dataset', - primary=primary) ) - - if isArchive is False: - download_from_query( query_item, target_output_filename ) - else: - target_output_path = os.path.join(appdata_path, filename) - download_from_query( query_item, target_output_path ) - if extra_data: - extra_files_path = ''.join( [ target_output_filename, 'files' ] ) - download_extra_data( extra_data, extra_files_path ) - - """ the following code handles archives and decompress them in a collection """ - if ( isArchive ): - db_key = "?" - archive_metadata = query_item.get( 'metadata', None ) - if archive_metadata is not None: - try: - db_key = archive_metadata.get( 'db_key' ) - except: - pass - walk_on_archive(target_output_path, check_ext, filename, appdata_path, db_key) - - return True - - -def set_up_config_values(json_params): - """ Parse json_params file and return a tuple of necessary configuration - values. - """ - datasource_params = json_params.get( 'param_dict' ) - dataset_url = datasource_params.get( 'URL' ) - output_filename = datasource_params.get( 'output1', None ) - output_data = json_params.get( 'output_data' ) - extra_files_path, file_name, ext, out_data_name, hda_id, dataset_id = \ - itemgetter('extra_files_path', 'file_name', 'ext', 'out_data_name', 'hda_id', 'dataset_id')(output_data[0]) - return (dataset_url, output_filename, - extra_files_path, file_name, - ext, out_data_name, - hda_id, dataset_id) - - -def download_from_json_data( options, args ): - """ Parse the returned JSON data and download files. Write metadata - to flat JSON file. - """ - output_base_path = options.path - appdata_path = options.appdata - if not os.path.exists(appdata_path): - os.makedirs(appdata_path) - - # read tool job configuration file and parse parameters we need - json_params = json.loads( open( options.json_param_file, 'r' ).read() ) - - dataset_url, output_filename, \ - extra_files_path, file_name, \ - ext, out_data_name, \ - hda_id, dataset_id = set_up_config_values(json_params) - # line separated JSON file to contain all dataset metadata - metadata_parameter_file = open( json_params['job_config']['TOOL_PROVIDED_JOB_METADATA_FILE'], 'wb' ) - - # get JSON response from data source - # TODO: make sure response is not enormous - query_params = json.loads(urllib.urlopen( dataset_url ).read()) - # download and write files - primary = False - #primary = True - # query_item, hda_id, output_base_path, dataset_id - for query_item in query_params: - if isinstance( query_item, list ): - # TODO: do something with the nested list as a collection - for query_subitem in query_item: - primary = download_files_and_write_metadata(query_subitem, json_params, output_base_path, - metadata_parameter_file, primary, appdata_path, options, args) - - elif isinstance( query_item, dict ): - primary = download_files_and_write_metadata(query_item, json_params, output_base_path, - metadata_parameter_file, primary, appdata_path, options, args) - metadata_parameter_file.close() - -def __main__(): - """ Read the JSON return from a data source. Parse each line and request - the data, download to "newfilepath", and write metadata. - - Schema - ------ - - [ {"url":"http://url_of_file", - "name":"encode WigData", - "extension":"wig", - "metadata":{"db_key":"hg19"}, - "extra_data":[ {"url":"http://url_of_ext_file", - "path":"rel/path/to/ext_file"} - ] - } - ] - - """ - # Parse the command line options - usage = "Usage: json_data_source_mod.py max_size --json_param_file filename [options]" - parser = optparse.OptionParser(usage = usage) - parser.add_option("-j", "--json_param_file", type="string", - action="store", dest="json_param_file", help="json schema return data") - parser.add_option("-p", "--path", type="string", - action="store", dest="path", help="new file path") - parser.add_option("-a", "--appdata", type="string", - action="store", dest="appdata", help="appdata folder name") - parser.add_option("-v", "--version", action="store_true", dest="version", - default=False, help="display version and exit") - - (options, args) = parser.parse_args() - if options.version: - print __version__ - else: - download_from_json_data( options, args ) - - -if __name__ == "__main__": __main__()