Mercurial > repos > damion > versioned_data
view versioned_data_cache_clear.py @ 1:5c5027485f7d draft
Uploaded correct file
author | damion |
---|---|
date | Sun, 09 Aug 2015 16:07:50 -0400 |
parents | |
children |
line wrap: on
line source
#!/usr/bin/python """ ****************************** versioned_data_cache_clear.py ****************************** Call this script directly to clear out all but the latest galaxy Versioned Data data library and server data store cached folder versions. SUGGEST RUNNING THIS UNDER GALAXY OR LESS PRIVILEGED USER, BUT the versioneddata_api_key file does need to be readable by the user. """ import vdb_retrieval import vdb_common import glob import os # Note that globals from vdb_retrieval can be referenced by prefixing with vdb_retrieval.XYZ # Note that this script uses the admin_api established in vdb_retrieval.py retrieval_obj = vdb_retrieval.VDBRetrieval() retrieval_obj.set_admin_api() retrieval_obj.user_api = retrieval_obj.admin_api retrieval_obj.set_datastores() workflow_keepers = [] #stack of Versioned Data library dataset_ids that if found in a workflow data input folder key name, can be saved; otherwise remove folder. library_folder_deletes = [] library_dataset_deletes = [] # Cycle through datastores, listing subfolders under each, sorted. # Permanently delete all but latest subfolder. for data_store in retrieval_obj.data_stores: spec_file_id = data_store['id'] # STEP 1: Determine data store type and location data_store_spec = retrieval_obj.admin_api.libraries.show_folder(retrieval_obj.library_id, spec_file_id) data_store_type = retrieval_obj.test_data_store_type(data_store_spec['name']) if not data_store_type in 'folder biomaj': # Folders are static - they don't do caching. base_folder_id = data_store_spec['folder_id'] ds_obj = retrieval_obj.get_data_store_gateway(data_store_type, spec_file_id) print #Cycle through library tree; have to look at the whole thing since there's no /[string]/* wildcard search: folders = retrieval_obj.get_library_folders(ds_obj.library_label_path) for ptr, folder in enumerate(folders): # Ignore folder that represents data store itself: if ptr == 0: print 'Data Store ::' + folder['name'] # Keep most recent cache item elif ptr == len(folders)-1: print 'Cached Version ::' + folder['name'] workflow_keepers.extend(folder['files']) # Drop version caches that are further in the past: else: print 'Clearing version cache:' + folder['name'] library_folder_deletes.extend(folder['id']) library_dataset_deletes.extend(folder['files']) # Now auto-clean versioned/ folders too? print "Server loc: " + ds_obj.data_store_path items = os.listdir(ds_obj.data_store_path) items = sorted(items, key=lambda el: vdb_common.natural_sort_key(el), reverse=True) count = 0 for name in items: # If it is a directory and it isn't the master or symlinked "current" one: # Add ability to skip sym-linked folders too? version_folder=os.path.join(ds_obj.data_store_path, name) if not name == 'master' \ and os.path.isdir(version_folder) \ and not os.path.islink(version_folder): count += 1 if count == 1: print "Keeping cache:" + name else: print "Dropping cache:" + name for root2, dirs2, files2 in os.walk(version_folder): for version_file in files2: full_path = os.path.join(root2, version_file) print "Removing " + full_path os.remove(full_path) #Not expecting any subfolders here. os.rmdir(version_folder) # Permanently delete specific data library datasets: for item in library_dataset_deletes: retrieval_obj.admin_api.libraries.delete_library_dataset(retrieval_obj.library_id, item['id'], purged=True) # Newer Bioblend API method for deleting galaxy library folders. # OLD Galaxy way possible: http DELETE request to {{url}}/api/folders/{{encoded_folder_id}}?key={{key}} if 'folders' in dir(retrieval_obj.admin_api): for folder in library_folder_deletes: retrieval_obj.admin_api.folders.delete(folder['id']) print workflow_keepers workflow_cache_folders = retrieval_obj.get_library_folders('/'+ vdb_retrieval.VDB_WORKFLOW_CACHE_FOLDER_NAME+'/') for folder in workflow_cache_folders: dataset_ids = folder['name'].split('_') #input dataset ids separated by underscore count = 0 for id in dataset_ids: if id in workflow_keepers: count += 1 # If every input dataset in workflow cache exists in library cache, then keep it. if count == len(dataset_ids): continue # We have one or more cached datasets to drop. print "Dropping workflow cache: " + folder['name'] for id in [item['id'] for item in folder['files']]: print id retrieval_obj.admin_api.libraries.delete_library_dataset(retrieval_obj.library_id, id, purged=True) # NOW DELETE WORKFLOW FOLDER. if 'folders' in dir(retrieval_obj.admin_api): retrieval_obj.admin_api.folders.delete(folder['id'])