0
|
1 #!/usr/bin/env python
|
|
2
|
|
3 """
|
|
4 Runs Lastz paired read alignment process
|
|
5 Written for Lastz v. 1.02.00.
|
|
6
|
|
7 # Author(s): based on various scripts written by Bob Harris (rsharris@bx.psu.edu),
|
|
8 # then tweaked to this form by Greg Von Kuster (greg@bx.psu.edu)
|
|
9
|
|
10 This tool takes the following input:
|
|
11 a. A collection of 454 paired end reads ( a fasta file )
|
|
12 b. A linker sequence ( a very small fasta file )
|
|
13 c. A reference genome ( nob, 2bit or fasta )
|
|
14
|
|
15 and uses the following process:
|
|
16 1. Split reads into mates: the input to this step is the read file XXX.fasta, and the output is three
|
|
17 files; XXX.short.fasta, XXX.long.fasta and XXX.mapping. The mapping file records the information necessary
|
|
18 to convert mate coordinates back into the original read, which is needed later in the process.
|
|
19
|
|
20 2. Align short mates to the reference: this runs lastz against every chromosome. The input is XXX.short.fasta
|
|
21 and the reference genome, and the output is a SAM file, XXX.short.sam.
|
|
22
|
|
23 3. Align long mates to the reference: this runs lastz against every chromosome. The input is XXX.long.fasta
|
|
24 and the reference genome, and the output is a SAM file, XXX.long.sam.
|
|
25
|
|
26 4. Combine, and convert mate coordinates back to read coordinates. The input is XXX.mapping, XXX.short.sam and
|
|
27 XXX.long.sam, and the output is XXX.sam.
|
|
28
|
|
29 usage: lastz_paired_reads_wrapper.py [options]
|
|
30 --ref_name: The reference name to change all output matches to
|
|
31 --ref_source: The reference is cached or from the history
|
|
32 --source_select: Use pre-set or cached reference file
|
|
33 --input1: The name of the reference file if using history or reference base name if using cached
|
|
34 --input2: The reads file to align
|
|
35 --input3: The sequencing linker file
|
|
36 --input4: The base quality score 454 file
|
|
37 --ref_sequences: The number of sequences in the reference file if using one from history
|
|
38 --output: The name of the output file
|
|
39 --lastz_seqs_file_dir: Directory of local lastz_seqs.loc file
|
|
40 """
|
|
41 import optparse, os, subprocess, shutil, sys, tempfile, time
|
|
42 from string import maketrans
|
|
43
|
|
44 from galaxy import eggs
|
|
45 import pkg_resources
|
|
46 pkg_resources.require( 'bx-python' )
|
|
47 from bx.seq.twobit import *
|
|
48 from bx.seq.fasta import FastaReader
|
|
49 from galaxy.util.bunch import Bunch
|
|
50 from galaxy.util import string_as_bool
|
|
51
|
|
52 # Column indexes for SAM required fields
|
|
53 SAM_QNAME_COLUMN = 0
|
|
54 SAM_FLAG_COLUMN = 1
|
|
55 SAM_RNAME_COLUMN = 2
|
|
56 SAM_POS_COLUMN = 3
|
|
57 SAM_MAPQ_COLUMN = 4
|
|
58 SAM_CIGAR_COLUMN = 5
|
|
59 SAM_MRNM_COLUMN = 6
|
|
60 SAM_MPOS_COLUMN = 7
|
|
61 SAM_ISIZE_COLUMN = 8
|
|
62 SAM_SEQ_COLUMN = 9
|
|
63 SAM_QUAL_COLUMN = 10
|
|
64 SAM_MIN_COLUMNS = 11
|
|
65 # SAM bit-encoded flags
|
|
66 BAM_FPAIRED = 1 # the read is paired in sequencing, no matter whether it is mapped in a pair
|
|
67 BAM_FPROPER_PAIR = 2 # the read is mapped in a proper pair
|
|
68 BAM_FUNMAP = 4 # the read itself is unmapped; conflictive with BAM_FPROPER_PAIR
|
|
69 BAM_FMUNMAP = 8 # the mate is unmapped
|
|
70 BAM_FREVERSE = 16 # the read is mapped to the reverse strand
|
|
71 BAM_FMREVERSE = 32 # the mate is mapped to the reverse strand
|
|
72 BAM_FREAD1 = 64 # this is read1
|
|
73 BAM_FREAD2 = 128 # this is read2
|
|
74 BAM_FSECONDARY = 256 # not primary alignment
|
|
75 BAM_FQCFAIL = 512 # QC failure
|
|
76 BAM_FDUP = 1024 # optical or PCR duplicate
|
|
77
|
|
78 # Keep track of all created temporary files so they can be deleted
|
|
79 global tmp_file_names
|
|
80 tmp_file_names = []
|
|
81 # The values in the skipped_lines dict are tuples consisting of:
|
|
82 # - the number of skipped lines for that error
|
|
83 # If not a sequence error:
|
|
84 # - the 1st line number on which the error was found
|
|
85 # - the text of the 1st line on which the error was found
|
|
86 # If a sequence error:
|
|
87 # - The number of the sequence in the file
|
|
88 # - the sequence name on which the error occurred
|
|
89 # We may need to improve dealing with file position and text as
|
|
90 # much of it comes from temporary files that are created from the
|
|
91 # inputs, and not the inputs themselves, so this could be confusing
|
|
92 # to the user.
|
|
93 global skipped_lines
|
|
94 skipped_lines = dict( bad_interval=( 0, 0, '' ),
|
|
95 inconsistent_read_lengths=( 0, 0, '' ),
|
|
96 inconsistent_reads=( 0, 0, '' ),
|
|
97 inconsistent_sizes=( 0, 0, '' ),
|
|
98 missing_mate=( 0, 0, '' ),
|
|
99 missing_quals=( 0, 0, '' ),
|
|
100 missing_seq=( 0, 0, '' ),
|
|
101 multiple_seqs=( 0, 0, '' ),
|
|
102 no_header=( 0, 0, '' ),
|
|
103 num_fields=( 0, 0, '' ),
|
|
104 reads_paired=( 0, 0, '' ),
|
|
105 sam_flag=( 0, 0, '' ),
|
|
106 sam_headers=( 0, 0, '' ),
|
|
107 sam_min_columns=( 0, 0, '' ),
|
|
108 two_mate_names=( 0, 0, '' ),
|
|
109 wrong_seq_len=( 0, 0, '' ) )
|
|
110 global total_skipped_lines
|
|
111 total_skipped_lines = 0
|
|
112
|
|
113 def stop_err( msg ):
|
|
114 sys.stderr.write( "%s" % msg )
|
|
115 sys.exit()
|
|
116
|
|
117 def skip_line( error_key, position, text ):
|
|
118 if not skipped_lines[ error_key ][2]:
|
|
119 skipped_lines[ error_key ][1] = position
|
|
120 skipped_lines[ error_key ][2] = text
|
|
121 skipped_lines[ error_key ][0] += 1
|
|
122 total_skipped_lines += 1
|
|
123
|
|
124 def get_tmp_file_name( dir=None, suffix=None ):
|
|
125 """
|
|
126 Return a unique temporary file name that can be managed. The
|
|
127 file must be manually removed after use.
|
|
128 """
|
|
129 if dir and suffix:
|
|
130 tmp_fd, tmp_name = tempfile.mkstemp( dir=dir, suffix=suffix )
|
|
131 elif dir:
|
|
132 tmp_fd, tmp_name = tempfile.mkstemp( dir=dir )
|
|
133 elif suffix:
|
|
134 tmp_fd, tmp_name = tempfile.mkstemp( suffix=suffix )
|
|
135 os.close( tmp_fd )
|
|
136 tmp_file_names.append( tmp_name )
|
|
137 return tmp_name
|
|
138
|
|
139 def run_command( command ):
|
|
140 proc = subprocess.Popen( args=command, shell=True, stderr=subprocess.PIPE, )
|
|
141 proc.wait()
|
|
142 stderr = proc.stderr.read()
|
|
143 proc.wait()
|
|
144 if stderr:
|
|
145 stop_err( stderr )
|
|
146
|
|
147 def split_paired_reads( input2, combined_linker_file_name ):
|
|
148 """
|
|
149 Given a fasta file of allegedly paired end reads ( input2 ), and a list of intervals
|
|
150 showing where the linker is on each read ( combined_linker_file_name ), split the reads into left and right
|
|
151 halves.
|
|
152
|
|
153 The input intervals look like this. Note that they may include multiple intervals for the same read
|
|
154 ( which should overlap ), and we use the union of them as the linker interval. Non-overlaps are
|
|
155 reported to the user, and those reads are not processed. Starts are origin zero.
|
|
156
|
|
157 #name strand start len size
|
|
158 FG3OYDA05FTEES + 219 42 283
|
|
159 FG3OYDA05FVOLL + 263 41 416
|
|
160 FG3OYDA05FFL7J + 81 42 421
|
|
161 FG3OYDA05FOQWE + 55 42 332
|
|
162 FG3OYDA05FV4DW + 297 42 388
|
|
163 FG3OYDA05FWAQV + 325 42 419
|
|
164 FG3OYDA05FVLGA + 90 42 367
|
|
165 FG3OYDA05FWJ71 + 58 42 276
|
|
166
|
|
167 The output gives each half-sequence on a separate line, like this. This allows easy sorting of the
|
|
168 sequences by length, after the fact.
|
|
169
|
|
170 219 FG3OYDA05FTEES_L TTTAGTTACACTTAACTCACTTCCATCCTCTAAATACGTGATTACCTTTC...
|
|
171 22 FG3OYDA05FTEES_R CCTTCCTTAAGTCCTAAAACTG
|
|
172 """
|
|
173 # Bob says these should be hard-coded.
|
|
174 seq_len_lower_threshold = 17
|
|
175 short_mate_cutoff = 50
|
|
176 # We need to pass the name of this file back to the caller.
|
|
177 tmp_mates_file_name = get_tmp_file_name( suffix='mates.txt' )
|
|
178 mates_file = file( tmp_mates_file_name, "w+b" )
|
|
179 # Read the linker intervals
|
|
180 combined_linker_file = file( combined_linker_file_name, "rb" )
|
|
181 read_to_linker_dict = {}
|
|
182 i = 0
|
|
183 for i, line in enumerate( combined_linker_file ):
|
|
184 line = line.strip()
|
|
185 if line.startswith( "#" ):
|
|
186 continue
|
|
187 if line.find( '#' ) >= 0:
|
|
188 line = line.split( "#", 1 )[0].rstrip()
|
|
189 fields = line.split()
|
|
190 if len( fields ) != 4:
|
|
191 skip_line( 'num_fields', i+1, line )
|
|
192 continue
|
|
193 name, start, length, size = fields
|
|
194 start = int( start )
|
|
195 length = int( length )
|
|
196 size = int( size )
|
|
197 end = start + length
|
|
198 if end > size:
|
|
199 skip_line[ 'bad_interval' ] += 1
|
|
200 continue
|
|
201 if name not in read_to_linker_dict:
|
|
202 read_to_linker_dict[ name ] = ( start, end, size )
|
|
203 continue
|
|
204 if read_to_linker_dict[ name ] == None:
|
|
205 # Read previously marked as non-overlapping intervals, so skip this sequence - see below
|
|
206 continue
|
|
207 ( s, e, sz ) = read_to_linker_dict[ name ]
|
|
208 if sz != size:
|
|
209 skip_line( 'inconsistent_sizes', i+1, name )
|
|
210 continue
|
|
211 if s > end or e < start:
|
|
212 # Non-overlapping intervals, so skip this sequence
|
|
213 read_to_linker_dict[ name ] = None
|
|
214 continue
|
|
215 read_to_linker_dict[ name ] = ( min( s, start ), max( e, end ), size )
|
|
216 combined_linker_file.close()
|
|
217 # We need to pass the name of this file back to the caller.
|
|
218 tmp_mates_mapping_file_name = get_tmp_file_name( suffix='mates.mapping' )
|
|
219 mates_mapping_file = file( tmp_mates_mapping_file_name, 'w+b' )
|
|
220 # Process the sequences
|
|
221 seqs = 0
|
|
222 fasta_reader = FastaReader( file( input2, 'rb' ) )
|
|
223 while True:
|
|
224 seq = fasta_reader.next()
|
|
225 if not seq:
|
|
226 break
|
|
227 seqs += 1
|
|
228 if seq.name not in read_to_linker_dict:
|
|
229 if seq.length > seq_len_lower_threshold:
|
|
230 mates_file.write( "%-3d %s %s\n" % ( seq.length, seq.name, seq.text ) )
|
|
231 read_to_linker_dict[ seq.name ] = ""
|
|
232 continue
|
|
233 if read_to_linker_dict[ seq.name ] == "":
|
|
234 skip_line( 'multiple_seqs', seqs, seq.name )
|
|
235 continue
|
|
236 if read_to_linker_dict[ seq.name ] == None:
|
|
237 # Read previously marked as non-overlapping intervals, so skip this sequence - see above
|
|
238 continue
|
|
239 ( start, end, size ) = read_to_linker_dict[ seq.name ]
|
|
240 if seq.length != size:
|
|
241 skip_line( 'wrong_seq_len', seqs, seq.name )
|
|
242 continue
|
|
243 left = seq.text[ :start ]
|
|
244 right = seq.text[ end: ]
|
|
245 left_is_small = len( left ) <= seq_len_lower_threshold
|
|
246 right_is_small = len( right ) <= seq_len_lower_threshold
|
|
247 if left_is_small and right_is_small:
|
|
248 continue
|
|
249 if not left_is_small:
|
|
250 mates_file.write( "%-3d %s %s\n" % ( len( left ), seq.name + "_L", left ) )
|
|
251 mates_mapping_file.write( "%s %s %s %s\n" % ( seq.name + "_L", seq.name, 0, size - start ) )
|
|
252 if not right_is_small:
|
|
253 mates_file.write( "%-3d %s %s\n" % ( len( right ), seq.name + "_R", right ) )
|
|
254 mates_mapping_file.write( "%s %s %s %s\n" % ( seq.name + "_R", seq.name, end, 0 ) )
|
|
255 read_to_linker_dict[ seq.name ] = ""
|
|
256 combined_linker_file.close()
|
|
257 mates_file.close()
|
|
258 mates_mapping_file.close()
|
|
259 # Create temporary files for short and long mates
|
|
260 tmp_mates_short_file_name = get_tmp_file_name( suffix='mates.short' )
|
|
261 tmp_mates_long_file_name = get_tmp_file_name( suffix='mates.long' )
|
|
262 tmp_mates_short = open( tmp_mates_short_file_name, 'w+b' )
|
|
263 tmp_mates_long = open( tmp_mates_long_file_name, 'w+b' )
|
|
264 i = 0
|
|
265 for i, line in enumerate( file( tmp_mates_file_name, 'rb' ) ):
|
|
266 fields = line.split()
|
|
267 seq_len = int( fields[0] )
|
|
268 seq_name = fields[1]
|
|
269 seq_text = fields[2]
|
|
270 if seq_len <= short_mate_cutoff:
|
|
271 tmp_mates_short.write( ">%s\n%s\n" % ( seq_name, seq_text ) )
|
|
272 else:
|
|
273 tmp_mates_long.write( ">%s\n%s\n" % ( seq_name, seq_text ) )
|
|
274 tmp_mates_short.close()
|
|
275 tmp_mates_long.close()
|
|
276 return tmp_mates_mapping_file_name, tmp_mates_file_name, tmp_mates_short_file_name, tmp_mates_long_file_name
|
|
277
|
|
278 def align_mates( input1, ref_source, ref_name, ref_sequences, tmp_mates_short_file_name, tmp_mates_long_file_name ):
|
|
279 tmp_align_file_names = []
|
|
280 if ref_source == 'history':
|
|
281 # Reference is a fasta dataset from the history
|
|
282 # Create temporary files to contain the output from lastz executions
|
|
283 tmp_short_file_name = get_tmp_file_name( suffix='short_out' )
|
|
284 tmp_align_file_names.append( tmp_short_file_name )
|
|
285 tmp_long_file_name = get_tmp_file_name( suffix='long_out' )
|
|
286 tmp_align_file_names.append( tmp_long_file_name )
|
|
287 seqs = 0
|
|
288 fasta_reader = FastaReader( open( input1 ) )
|
|
289 while True:
|
|
290 # Read the next sequence from the reference dataset. Note that if the reference contains
|
|
291 # a small number of chromosomes this loop is ok, but in many cases the genome has a bunch
|
|
292 # of small straggler scaffolds and contigs and it is a computational waste to do each one
|
|
293 # of these in its own run. There is an I/O down side to running by subsets (even if they are
|
|
294 # one sequence per subset), compared to splitting the reference into sizes of 250 mb. With
|
|
295 # the subset action, lastz still has to read and parse the entire file for every run (this
|
|
296 # is true for fasta, but for .2bit files it can access each sequence directly within the file,
|
|
297 # so the overhead is minimal).
|
|
298 """
|
|
299 :> output_file (this creates the output file, empty)
|
|
300 while there are more sequences to align
|
|
301 find the next sequences that add up to 250M, put their names in farf.names
|
|
302 lastz ${refFile}[subset=farf.names][multi][unmask] ${matesPath}/${matesFile} ...
|
|
303 >> output_file
|
|
304 """
|
|
305 seq = fasta_reader.next()
|
|
306 if not seq:
|
|
307 break
|
|
308 seqs += 1
|
|
309 # Create a temporary file to contain the current sequence as input to lastz.
|
|
310 # We're doing this a bit differently here since we could be generating a huge
|
|
311 # number of temporary files.
|
|
312 tmp_in_fd, tmp_in_file_name = tempfile.mkstemp( suffix='seq_%d_in' % seqs )
|
|
313 tmp_in_file = os.fdopen( tmp_in_fd, 'w+b' )
|
|
314 tmp_in_file.write( '>%s\n%s\n' % ( seq.name, seq.text ) )
|
|
315 tmp_in_file.close()
|
|
316 # Align short mates
|
|
317 command = 'lastz %s[unmask]%s %s ' % ( tmp_in_file_name, ref_name, tmp_mates_short_file_name )
|
|
318 command += 'Z=1 --seed=1111111011111 --notrans --maxwordcount=90% --match=1,3 O=1 E=3 X=15 K=10 Y=12 L=18 --ambiguousn --noytrim --identity=95 --coverage=80 --continuity=95 --format=softsam- '
|
|
319 command += '>> %s' % tmp_short_file_name
|
|
320 run_command( command )
|
|
321 # Align long mates
|
|
322 command = 'lastz %s[unmask]%s %s ' % ( tmp_in_file_name, ref_name, tmp_mates_long_file_name )
|
|
323 command += 'Z=15 W=13 --notrans --exact=18 --maxwordcount=90% --match=1,3 O=1 E=3 Y=10 L=18 --ambiguousn --noytrim --identity=95 --coverage=90 --continuity=95 --format=softsam- '
|
|
324 command += '>> %s' % tmp_long_file_name
|
|
325 run_command( command )
|
|
326 # Remove the temporary file that contains the current sequence
|
|
327 os.remove( tmp_in_file_name )
|
|
328 else:
|
|
329 # Reference is a locally cached 2bit file, split lastz calls across number of chroms in 2bit file
|
|
330 tbf = TwoBitFile( open( input1, 'rb' ) )
|
|
331 for chrom in tbf.keys():
|
|
332 # Align short mates
|
|
333 tmp_short_file_name = get_tmp_file_name( suffix='short_vs_%s' % chrom )
|
|
334 tmp_align_file_names.append( tmp_short_file_name )
|
|
335 command = 'lastz %s/%s[unmask]%s %s ' % ( input1, chrom, ref_name, tmp_mates_short_file_name )
|
|
336 command += 'Z=1 --seed=1111111011111 --notrans --maxwordcount=90% --match=1,3 O=1 E=3 X=15 K=10 Y=12 L=18 --ambiguousn --noytrim --identity=95 --coverage=80 --continuity=95 --format=softsam- '
|
|
337 command += '> %s' % tmp_short_file_name
|
|
338 run_command( command )
|
|
339 # Align long mates
|
|
340 tmp_long_file_name = get_tmp_file_name( suffix='long_vs_%s' % chrom )
|
|
341 tmp_align_file_names.append( tmp_long_file_name )
|
|
342 command = 'lastz %s/%s[unmask]%s %s ' % ( input1, chrom, ref_name, tmp_mates_long_file_name )
|
|
343 command += 'Z=15 W=13 --notrans --exact=18 --maxwordcount=90% --match=1,3 O=1 E=3 Y=10 L=18 --ambiguousn --noytrim --identity=95 --coverage=90 --continuity=95 --format=softsam- '
|
|
344 command += '> %s' % tmp_long_file_name
|
|
345 run_command( command )
|
|
346 return tmp_align_file_names
|
|
347
|
|
348 def paired_mate_unmapper( input2, input4, tmp_mates_mapping_file_name, tmp_align_file_name_list, output ):
|
|
349 """
|
|
350 Given a SAM file corresponding to alignments of *subsegments* of paired 'reads' to a reference sequence,
|
|
351 convert the positions on the subsegments to positions on the reads. Also (optionally) add quality values.
|
|
352
|
|
353 The input file is in SAM format, as shown below. Each line represents the alignment of a part of a read
|
|
354 to a reference sequence. Read pairs are indicated by suffixes in their names. Normally, the suffixes _L
|
|
355 and _R indicate the left and right mates of reads (this can be overridden with the --left and --right
|
|
356 options). Reads that were not mates have no suffix.
|
|
357
|
|
358 (SAM header lines omitted)
|
|
359 F2YP0BU02G7LK5_R 16 chr21 15557360 255 40M * 0 0 ATTTTATTCTCTTTGAAGCAATTGTGAATGGGAGTTTACT *
|
|
360 F2YP0BU02HXV58_L 16 chr21 15952091 255 40M6S * 0 0 GCAAATTGTGCTGCTTTAAACATGCGTGTGCAAGTATCTTtttcat *
|
|
361 F2YP0BU02HREML_R 0 chr21 16386077 255 33M5S * 0 0 CCAAAGTTCTGGGATTACAGGCGTGAGCCATCGcgccc *
|
|
362 F2YP0BU02IOF1F_L 0 chr21 17567321 255 7S28M * 0 0 taaagagAAGAATTCTCAACCCAGAATTTCATATC *
|
|
363 F2YP0BU02IKX84_R 16 chr21 18491628 255 22M1D18M9S * 0 0 GTCTCTACCAAAAAATACAAAAATTAGCCGGGCGTGGTGGcatgtctgt *
|
|
364 F2YP0BU02GW5VA_L 16 chr21 20255344 255 6S32M * 0 0 caagaaCAAACACATTCAAAAGCTAGTAGAAGGCAAGA *
|
|
365 F2YP0BU02JIMJ4_R 0 chr21 22383051 255 19M * 0 0 CCCTTTATCATTTTTTATT *
|
|
366 F2YP0BU02IXZGF_L 16 chr21 23094798 255 13M1I18M * 0 0 GCAAGCTCCACTTCCCGGGTTCACGCCATTCT *
|
|
367 F2YP0BU02IODR5_L 0 chr21 30935325 255 37M * 0 0 GAAATAAAGGGTATTCAATTAGGAAAAGAGGAAGTCA *
|
|
368 F2YP0BU02IMZBL_L 16 chr21 31603486 255 28M1D1M * 0 0 ATACAAAAATTAGCCGGGCACAGTGGCAG *
|
|
369 F2YP0BU02JA9PR_L 16 chr21 31677159 255 23M * 0 0 CACACCTGTAACCCCAGCACTTT *
|
|
370 F2YP0BU02HKC61_R 0 chr21 31678718 255 40M * 0 0 CACTGCACTCCAGCCTGGGTGACAAAGCAAGACTCTGTCT *
|
|
371 F2YP0BU02HKC61_R 0 chr21 31678718 255 40M * 0 0 CACTGCACTCCAGCCTGGGTGACAAAGCAAGACTCTGTCT *
|
|
372 F2YP0BU02HVA88 16 chr21 31703558 255 1M1D35M8S * 0 0 TGGGATTACAGGCGTGAGCTACCACACCCAGCCAGAgttcaaat *
|
|
373 F2YP0BU02JDCF1_L 0 chr21 31816600 255 38M * 0 0 AGGAGAATCGCTTGAACCCAGGAGGCAGAGGTTGCGGT *
|
|
374 F2YP0BU02GZ1GO_R 0 chr21 33360122 255 6S38M * 0 0 cctagaCTTCACACACACACACACACACACACACACACACACAC *
|
|
375 F2YP0BU02FX387_L 16 chr22 14786201 255 26M * 0 0 TGGATGAAGCTGGAAACCATCATTCT *
|
|
376 F2YP0BU02IF2NE_R 0 chr22 16960842 255 40M10S * 0 0 TGGCATGCACCTGTAGTCTCAGCTACTTGGGAGGCTGAGGtgggaggatc *
|
|
377 F2YP0BU02F4TVA 0 chr22 19200522 255 49M * 0 0 CCTGGGAGGCGGAGGTTGCAGTGAGCCGAGATCACGCCATTGCACTCCA *
|
|
378 F2YP0BU02HKC61_R 16 chr22 29516998 255 8S32M * 0 0 agacagagTCTTGCTTTGTCACCCAGGCTGGAGTGCAGTG *
|
|
379 F2YP0BU02FS4EM_R 0 chr22 30159364 255 29M * 0 0 CTCCTGCCTCAGCCTCCCGAGTAGTTGGG *
|
|
380 F2YP0BU02G197P_L 0 chr22 32044496 255 40M10S * 0 0 TTGTTGGACATTTGGGTTGGTTCCAAGTCTTTGCTATTGTgaataatgcc *
|
|
381 F2YP0BU02FIING 16 chr22 45959944 255 3M1I11M1I26M * 0 0 AGCTATGGTACTGGCTATGAAAGCAGACACATAGACCAATGG *
|
|
382 F2YP0BU02GUB9L_L 16 chr22 49198404 255 16M1I20M * 0 0 CACCACGCTCGGCTAATTTTTGTATTTTTAGTAGAGA *
|
|
383
|
|
384 The user must provide a mapping file (which might better be called an unmapping file). This file is usually
|
|
385 created by split_paired_reads, and tells us how to map the subsegments back to original coordinates in a single
|
|
386 read (this means the left and right mates were part of a single read). The mapping file contains four columns.
|
|
387 The first two give the mates's name (including the suffix) and the read name. The last two columns describe how
|
|
388 much of the full original sequence is missing from the mate. For example, in the read below, the left mate is
|
|
389 missing 63 on the right (42 for the linker and 21 for the right half). The right mate is missing 339 on the left.
|
|
390
|
|
391 left half: TTTCAACATATGCAAATCAATAAATGTAATCCAGCATATAAACAGAACCA
|
|
392 AAGACAAAAACCACATGATTATCTCAATAGATGCAGAAAAGGCCTTCGGC
|
|
393 AAAATTCAACAAAACTCCATGCTAAAACTCTCAATAAGGTATTGATGGGA
|
|
394 CATGCCGCATAATAATAAGACATATCTATGACAAACCCACAGCCAATATC
|
|
395 ATGCTGAATGCACAAAAATTGGAAGCATTCCCTTTGAAAACTGGCACAAG
|
|
396 ACTGGGATGCCCTCTCTCACAACTCCTATTCAACATAGTGTTGGAAG
|
|
397 linker: CGTAATAACTTCGTATAGCATACATTATACGAAGTCATACGA
|
|
398 right half: CTCCTGCCTCAGCCTCCCGAG
|
|
399
|
|
400 mate_name read_name offset_to_start offset_from_end
|
|
401 F2YP0BU02FS4EM_L F2YP0BU02FS4EM 0 71
|
|
402 F2YP0BU02FS4EM_R F2YP0BU02FS4EM 339 0
|
|
403
|
|
404 The user can also specify a quality scores file, which should look something like this. Quality values are presumed
|
|
405 to be PHRED scores, written in space-delimited decimal.
|
|
406
|
|
407 >F2YP0BU02FS4EM
|
|
408 38 38 38 40 40 40 40 40 40 40 40 40 40 39 39 39 40 40 40 40 38 21 21 21 40
|
|
409 40 40 40 40 40 40 40 40 40 40 40 40 40 40 39 39 39 40 40 40 40 40 40 40 33
|
|
410 32 32 40 40 40 21 21 18 18 21 34 34 31 40 40 40 40 40 40 40 40 40 40 40 40
|
|
411 40 40 40 40 40 40 40 40 40 40 40 32 32 32 32 40 40 40 40 40 40 40 34 34 35
|
|
412 31 31 28 28 33 33 33 36 36 36 17 17 17 19 26 36 36 36 40 40 40 40 40 33 34
|
|
413 34 34 39 39 39 40 40 40 40 40 33 33 34 34 40 40 40 40 40 40 40 39 39 39 40
|
|
414 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40
|
|
415 40 40 40 40 40 40 40 39 39 39 39 39 39 40 40 40 39 39 39 40 40 40 40 40 40
|
|
416 40 40 40 40 40 40 40 40 40 40 40 40 40 26 26 26 26 26 40 40 38 38 37 35 33
|
|
417 36 40 19 17 17 17 17 19 19 23 30 20 20 20 23 35 40 36 36 36 36 36 36 36 36
|
|
418 39 40 34 20 27 27 35 39 40 37 40 40 40 40 40 40 40 40 40 40 34 34 35 39 40
|
|
419 40 40 40 40 40 40 39 39 39 40 40 40 40 36 36 32 32 28 28 29 30 36 40 30 26
|
|
420 26 26 34 39 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 39 39 39
|
|
421 40 39 35 34 34 40 40 40 40 30 30 30 35 40 40 40 40 40 39 39 36 40 40 40 40
|
|
422 39 39 39 39 30 30 28 35 35 39 40 40 40 40 40 35 35 35
|
|
423 >F2YP0BU02G197P
|
|
424 40 40 40 40 40 40 40 40 40 40 39 39 39 39 39 39 40 40 40 40 40 40 40 40 40
|
|
425 40 40 40 40 26 26 26 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40
|
|
426 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40
|
|
427 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40
|
|
428 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 34 34 34 40 40
|
|
429 40 40 40 40 40 40 39 39 39 40 40 40 40 40 40 40 40 40 40 39 39 39 40 40 40
|
|
430 40 40 40 40 40 40 40 34 34 34 34 40 40 40 40 34 34 34 34 40 40 40 40 40 40
|
|
431 40 40 40 40 40 39 39 39 34 34 34 34 40 40 40 40 39 39 25 25 26 39 40 40 40
|
|
432 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40
|
|
433 33 33 33 33 40 35 21 21 21 30 38 40 40 40 40 40 40 40 40 35 35 30 30 30 40
|
|
434 40 40 39 39 39 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40 40
|
|
435 40 40 40 40 40 40 40 40 40 40 40 40 39 39 39 40 40 40 40 40 40 40 40 40 40
|
|
436 40 40 40 39 39 39 40 40
|
|
437 >F2YP0BU02FIING
|
|
438 32 32 32 25 25 25 25 24 25 30 31 30 27 27 27 28 28 21 19 19 13 13 13 14 19
|
|
439 19 17 19 16 16 25 28 22 21 17 17 18 25 24 25 25 25
|
|
440
|
|
441 The output file is also SAM:
|
|
442
|
|
443 (SAM header lines omitted)
|
|
444 F2YP0BU02G7LK5 81 chr21 15557360 255 40M303H * 0 0 ATTTTATTCTCTTTGAAGCAATTGTGAATGGGAGTTTACT D>>>>IIIIIIHHG???IIIIIIIIIHHHFFEIH999HII
|
|
445 F2YP0BU02HXV58 145 chr21 15952091 255 226H40M6S * 0 0 GCAAATTGTGCTGCTTTAAACATGCGTGTGCAAGTATCTTtttcat AA===DDDDAAAAD???:::ABBBBBAAA:888ECF;F>>>?8??@
|
|
446 F2YP0BU02HREML 65 chr21 16386077 255 320H33M5S * 0 0 CCAAAGTTCTGGGATTACAGGCGTGAGCCATCGcgccc HH???HHIIIHFHIIIIIIICDDHHIIIIIIHHHHHHH
|
|
447 F2YP0BU02IOF1F 129 chr21 17567321 255 7S28M409H * 0 0 taaagagAAGAATTCTCAACCCAGAATTTCATATC 4100<<A>4113:<EFGGGFFFHHHHHHDFFFFED
|
|
448 F2YP0BU02IKX84 81 chr21 18491628 255 22M1D18M9S341H * 0 0 GTCTCTACCAAAAAATACAAAAATTAGCCGGGCGTGGTGGcatgtctgt ;;;=7@.55------?2?11112GGB=CCCCDIIIIIIIIIHHHHHHII
|
|
449 F2YP0BU02GW5VA 145 chr21 20255344 255 286H6S32M * 0 0 caagaaCAAACACATTCAAAAGCTAGTAGAAGGCAAGA IIIIIIIHHHIIIIIIICCCCIIIIIIIIIIIIIIIII
|
|
450 F2YP0BU02JIMJ4 65 chr21 22383051 255 208H19M * 0 0 CCCTTTATCATTTTTTATT 555544E?GE113344I22
|
|
451 F2YP0BU02IXZGF 145 chr21 23094798 255 291H13M1I18M * 0 0 GCAAGCTCCACTTCCCGGGTTCACGCCATTCT IIIIIIIIIIIGG;;;GGHIIIIIGGGIIIII
|
|
452 F2YP0BU02IODR5 129 chr21 30935325 255 37M154H * 0 0 GAAATAAAGGGTATTCAATTAGGAAAAGAGGAAGTCA 6...7/--..,30;9<<>@BFFFAAAAHIIIIIH@@@
|
|
453 F2YP0BU02IMZBL 145 chr21 31603486 255 342H28M1D1M * 0 0 ATACAAAAATTAGCCGGGCACAGTGGCAG BB1552222<<>9==8;;?AA=??A???A
|
|
454 F2YP0BU02JA9PR 145 chr21 31677159 255 229H23M * 0 0 CACACCTGTAACCCCAGCACTTT IIIIIIIIIIICCCCIIIIIHHH
|
|
455 F2YP0BU02HKC61 65 chr21 31678718 255 300H40M * 0 0 CACTGCACTCCAGCCTGGGTGACAAAGCAAGACTCTGTCT AA@BD:::==AAA@A?8888:<90004<>>?><<<<4442
|
|
456 F2YP0BU02HKC61 65 chr21 31678718 255 300H40M * 0 0 CACTGCACTCCAGCCTGGGTGACAAAGCAAGACTCTGTCT AA@BD:::==AAA@A?8888:<90004<>>?><<<<4442
|
|
457 F2YP0BU02HVA88 16 chr21 31703558 255 1M1D35M8S * 0 0 TGGGATTACAGGCGTGAGCTACCACACCCAGCCAGAgttcaaat >8888DFFHHGFHHHH@@?@?DDC96666HIIIFFFFFFFFFFF
|
|
458 F2YP0BU02JDCF1 129 chr21 31816600 255 38M103H * 0 0 AGGAGAATCGCTTGAACCCAGGAGGCAGAGGTTGCGGT IIIIIIIIIIIHHHIIHHHIIIIIIIIIIIIIIIIIII
|
|
459 F2YP0BU02GZ1GO 65 chr21 33360122 255 76H6S38M * 0 0 cctagaCTTCACACACACACACACACACACACACACACACACAC BBBBD?:688CFFFFFFFFFFFFFFFFFFFFFFFFFFDDBBB51
|
|
460 F2YP0BU02FX387 145 chr22 14786201 255 201H26M * 0 0 TGGATGAAGCTGGAAACCATCATTCT IIHHHHHHHHHHHHHFFFFFFFFFFF
|
|
461 F2YP0BU02IF2NE 65 chr22 16960842 255 209H40M10S * 0 0 TGGCATGCACCTGTAGTCTCAGCTACTTGGGAGGCTGAGGtgggaggatc BAAADDDDFDDDDDDBBA889<A?4444000@<>AA?9444;;8>77<7-
|
|
462 F2YP0BU02F4TVA 0 chr22 19200522 255 49M * 0 0 CCTGGGAGGCGGAGGTTGCAGTGAGCCGAGATCACGCCATTGCACTCCA FFF???FFFFFIIIIIIIIIIIIIIIIIIIIIIIHHIIFHFFFGDDB=5
|
|
463 F2YP0BU02HKC61 81 chr22 29516998 255 8S32M300H * 0 0 agacagagTCTTGCTTTGTCACCCAGGCTGGAGTGCAGTG 2444<<<<>?>><40009<:8888?A@AAA==:::DB@AA
|
|
464 F2YP0BU02FS4EM 65 chr22 30159364 255 339H29M * 0 0 CTCCTGCCTCAGCCTCCCGAGTAGTTGGG IIIIHHEIIIIHHHH??=DDHIIIIIDDD
|
|
465 F2YP0BU02G197P 129 chr22 32044496 255 40M10S258H * 0 0 TTGTTGGACATTTGGGTTGGTTCCAAGTCTTTGCTATTGTgaataatgcc IIIIIIIIIIHHHHHHIIIIIIIIIIIII;;;IIIIIIIIIIIIIIIIII
|
|
466 F2YP0BU02FIING 16 chr22 45959944 255 3M1I11M1I26M * 0 0 AGCTATGGTACTGGCTATGAAAGCAGACACATAGACCAATGG :::9:32267=:114244/...446==<<<?@?:9::::AAA
|
|
467 F2YP0BU02GUB9L 145 chr22 49198404 255 176H16M1I20M * 0 0 CACCACGCTCGGCTAATTTTTGTATTTTTAGTAGAGA IIIIIIIIIHAAC;<</////@4F5778;IIIIIIII
|
|
468
|
|
469 """
|
|
470 left_suffix = "_L"
|
|
471 right_suffix = "_R"
|
|
472 # Read the mapping
|
|
473 mate_to_read_dict = {}
|
|
474 i = 0
|
|
475 for i, line in enumerate( file( tmp_mates_mapping_file_name, 'rb' ) ):
|
|
476 line = line.strip()
|
|
477 if not line.startswith( "#" ):
|
|
478 fields = line.split()
|
|
479 if len( fields ) != 4:
|
|
480 skip_line( "num_fields", i+1, line )
|
|
481 continue
|
|
482 mate_name, read_name, s_offset, e_offset = fields
|
|
483 if mate_name in mate_to_read_dict:
|
|
484 skip_line( 'two_mate_names', i+1, mate_name )
|
|
485 continue
|
|
486 mate_to_read_dict[ mate_name ] = ( read_name, int( s_offset ), int( e_offset ) )
|
|
487 # Read sequence data
|
|
488 read_to_nucs_dict = {}
|
|
489 seqs = 0
|
|
490 fasta_reader = FastaReader( file( input2, 'rb' ) )
|
|
491 while True:
|
|
492 seq = fasta_reader.next()
|
|
493 if not seq:
|
|
494 break
|
|
495 seqs += 1
|
|
496 seq_text_upper = seq.text.upper()
|
|
497 if seq.name in read_to_nucs_dict:
|
|
498 if seq_text_upper != read_to_nucs_dict[ seq.name ]:
|
|
499 skip_line( 'inconsistent_reads', seqs, seq.name )
|
|
500 continue
|
|
501 read_to_nucs_dict[ seq.name ] = seq_text_upper
|
|
502 # Read quality data
|
|
503 def quality_sequences( f ):
|
|
504 seq_name = None
|
|
505 seq_quals = None
|
|
506 line_number = 0
|
|
507 for line in f:
|
|
508 line_number += 1
|
|
509 line = line.strip()
|
|
510 if line.startswith( ">" ):
|
|
511 if seq_name != None:
|
|
512 yield ( seq_name, seq_quals, seq_line )
|
|
513 seq_name = sequence_name( line )
|
|
514 seq_line = line_number
|
|
515 seq_quals = []
|
|
516 elif seq_name is None:
|
|
517 skip_line( 'no_header', line_number, line )
|
|
518 continue
|
|
519 else:
|
|
520 seq_quals += [ int( q ) for q in line.split() ]
|
|
521 if seq_name is not None:
|
|
522 yield ( seq_name, seq_quals, seq_line )
|
|
523 def sequence_name( s ):
|
|
524 s = s[ 1: ].strip()
|
|
525 if not s:
|
|
526 return ""
|
|
527 else:
|
|
528 return s.split()[ 0 ]
|
|
529 read_to_quals_dict = {}
|
|
530 # TODO: should we use Dan's fastaNamedReader here?
|
|
531 for seq_name, quals, line_number in quality_sequences( file( input4 ) ):
|
|
532 quals = samify_phred_scores( quals )
|
|
533 if seq_name in read_to_quals_dict:
|
|
534 if quals != read_to_quals_dict[ seq_name ]:
|
|
535 skip_line( 'inconsistent_reads', line_number, seq_name )
|
|
536 continue
|
|
537 if len( quals ) != len( read_to_nucs_dict[ seq_name ] ):
|
|
538 skip_line( 'inconsistent_read_lengths', line_number, seq_name )
|
|
539 continue
|
|
540 read_to_quals_dict[ seq_name ] = quals
|
|
541 # process the SAM file
|
|
542 tmp_align_file_names = ' '.join( tmp_align_file_name_list )
|
|
543 combined_chrom_file_name = get_tmp_file_name( suffix='combined_chrom' )
|
|
544 command = 'cat %s | grep -v "^@" | sort -k 1 > %s' % ( tmp_align_file_names, combined_chrom_file_name )
|
|
545 run_command( command )
|
|
546 fout = file( output, 'w+b' )
|
|
547 has_non_header = False
|
|
548 i = 0
|
|
549 for i, line in enumerate( file( combined_chrom_file_name, 'rb' ) ):
|
|
550 line = line.strip()
|
|
551 if line.startswith( "@" ):
|
|
552 if has_non_header:
|
|
553 skip_line( 'sam_headers', i+1, line )
|
|
554 continue
|
|
555 fout.write( "%s\n" % line )
|
|
556 continue
|
|
557 has_non_header = True
|
|
558 fields = line.split()
|
|
559 num_fields = len( fields )
|
|
560 if num_fields < SAM_MIN_COLUMNS:
|
|
561 skip_line( 'sam_min_columns', i+1, line )
|
|
562 continue
|
|
563 # Set flags for mates
|
|
564 try:
|
|
565 flag = int( fields[ SAM_FLAG_COLUMN ] )
|
|
566 except ValueError:
|
|
567 skip_line( 'sam_flag', i+1, line )
|
|
568 continue
|
|
569 if not( flag & ( BAM_FPAIRED + BAM_FREAD1 + BAM_FREAD2 ) == 0 ):
|
|
570 skip_line( 'reads_paired', i+1, line )
|
|
571 continue
|
|
572 mate_name = fields[ SAM_QNAME_COLUMN ]
|
|
573 unmap_it = False
|
|
574 half = None
|
|
575 if mate_name.endswith( left_suffix ):
|
|
576 flag += BAM_FPAIRED + BAM_FREAD2
|
|
577 fields[ SAM_FLAG_COLUMN ] = "%d" % flag
|
|
578 unmap_it = True
|
|
579 half = "L"
|
|
580 elif mate_name.endswith( right_suffix ):
|
|
581 flag += BAM_FPAIRED + BAM_FREAD1
|
|
582 fields[ SAM_FLAG_COLUMN ] = "%d" % flag
|
|
583 unmap_it = True
|
|
584 half = "R"
|
|
585 on_plus_strand = ( flag & BAM_FREVERSE == 0 )
|
|
586 # Convert position from mate to read by adding clipping to cigar
|
|
587 if not unmap_it:
|
|
588 read_name = mate_name
|
|
589 else:
|
|
590 try:
|
|
591 read_name, s_offset, e_offset = mate_to_read_dict[ mate_name ]
|
|
592 except KeyError:
|
|
593 skip_line( 'missing_mate', i+1, mate_name )
|
|
594 continue
|
|
595 cigar = fields[ SAM_CIGAR_COLUMN ]
|
|
596 cigar_prefix = None
|
|
597 cigar_suffix = None
|
|
598 if half == "L":
|
|
599 if on_plus_strand:
|
|
600 if s_offset > 0:
|
|
601 cigar_prefix = ( s_offset, "S" )
|
|
602 if e_offset > 0:
|
|
603 cigar_suffix = ( e_offset, "H" )
|
|
604 else:
|
|
605 if e_offset > 0:
|
|
606 cigar_prefix = ( e_offset, "H" )
|
|
607 if s_offset > 0:
|
|
608 cigar_suffix = ( s_offset, "S" )
|
|
609 elif half == "R":
|
|
610 if on_plus_strand:
|
|
611 if s_offset > 0:
|
|
612 cigar_prefix = ( s_offset, "H" )
|
|
613 if e_offset > 0:
|
|
614 cigar_suffix = ( e_offset, "S" )
|
|
615 else:
|
|
616 if e_offset > 0:
|
|
617 cigar_prefix = ( e_offset, "S" )
|
|
618 if s_offset > 0:
|
|
619 cigar_suffix = ( s_offset, "H" )
|
|
620 else:
|
|
621 if on_plus_strand:
|
|
622 if s_offset > 0:
|
|
623 cigar_prefix = ( s_offset, "S" )
|
|
624 if e_offset > 0:
|
|
625 cigar_suffix = ( e_offset, "S" )
|
|
626 else:
|
|
627 if e_offset > 0:
|
|
628 cigar_prefix = ( e_offset, "S" )
|
|
629 if s_offset > 0:
|
|
630 cigar_suffix = ( s_offset, "S" )
|
|
631 if cigar_prefix != None:
|
|
632 count, op = cigar_prefix
|
|
633 cigar = prefix_cigar( "%d%s" % ( count, op ), cigar )
|
|
634 if op == "S":
|
|
635 refPos = int( fields[ SAM_POS_COLUMN ] ) - count
|
|
636 fields[ SAM_POS_COLUMN ] = "%d" % refPos
|
|
637 if cigar_suffix != None:
|
|
638 count, op = cigar_suffix
|
|
639 cigar = suffix_cigar( cigar,"%d%s" % ( count, op) )
|
|
640 fields[ SAM_QNAME_COLUMN ] = read_name
|
|
641 fields[ SAM_CIGAR_COLUMN ] = cigar
|
|
642 # Fetch sequence and quality values, and flip/clip them
|
|
643 if read_name not in read_to_nucs_dict:
|
|
644 skip_line( 'missing_seq', i+1, read_name )
|
|
645 continue
|
|
646 nucs = read_to_nucs_dict[ read_name ]
|
|
647 if not on_plus_strand:
|
|
648 nucs = reverse_complement( nucs )
|
|
649 quals = None
|
|
650 if read_to_quals_dict != None:
|
|
651 if read_name not in read_to_quals_dict:
|
|
652 skip_line( 'missing_quals', i+1, read_name )
|
|
653 continue
|
|
654 quals = read_to_quals_dict[ read_name ]
|
|
655 if not on_plus_strand:
|
|
656 quals = reverse_string( quals )
|
|
657 cigar = split_cigar( fields[ SAM_CIGAR_COLUMN ] )
|
|
658 nucs, quals = clip_for_cigar( cigar, nucs, quals )
|
|
659 fields[ SAM_SEQ_COLUMN ] = nucs
|
|
660 if quals != None:
|
|
661 fields[ SAM_QUAL_COLUMN ] = quals
|
|
662 # Output the line
|
|
663 fout.write( "%s\n" % "\t".join( fields ) )
|
|
664 fout.close()
|
|
665
|
|
666 def prefix_cigar( prefix, cigar ):
|
|
667 ix = 0
|
|
668 while cigar[ ix ].isdigit():
|
|
669 ix += 1
|
|
670 if cigar[ ix ] != prefix[ -1 ]:
|
|
671 return prefix + cigar
|
|
672 count = int( prefix[ :-1 ] ) + int( cigar[ :ix ] )
|
|
673 return "%d%s%s" % ( count, prefix[ -1 ], cigar[ ix+1: ] )
|
|
674
|
|
675 def suffix_cigar( cigar, suffix ):
|
|
676 if cigar[ -1 ] != suffix[ -1 ]:
|
|
677 return cigar + suffix
|
|
678 ix = len( cigar ) - 2
|
|
679 while cigar[ix].isdigit():
|
|
680 ix -= 1
|
|
681 ix += 1
|
|
682 count = int( cigar[ ix:-1 ] ) + int( suffix[ :-1 ] )
|
|
683 return "%s%d%s" % ( cigar[ :ix ], count, suffix[ -1 ] )
|
|
684
|
|
685 def split_cigar( text ):
|
|
686 fields = []
|
|
687 field = []
|
|
688 for ch in text:
|
|
689 if ch not in "MIDHS":
|
|
690 field += ch
|
|
691 continue
|
|
692 if field == []:
|
|
693 raise ValueError
|
|
694 fields += [ ( int( "".join( field ) ), ch ) ]
|
|
695 field = []
|
|
696 if field != []:
|
|
697 raise ValueError
|
|
698 return fields
|
|
699
|
|
700 def clip_for_cigar( cigar, nucs, quals ):
|
|
701 # Hard clip prefix
|
|
702 count, op = cigar[0]
|
|
703 if op == "H":
|
|
704 nucs = nucs[ count: ]
|
|
705 if quals != None:
|
|
706 quals = quals[ count: ]
|
|
707 count, op = cigar[ 1 ]
|
|
708 # Soft clip prefix
|
|
709 if op == "S":
|
|
710 nucs = nucs[ :count ].lower() + nucs[ count: ]
|
|
711 # Hard clip suffix
|
|
712 count,op = cigar[ -1 ]
|
|
713 if op == "H":
|
|
714 nucs = nucs[ :-count ]
|
|
715 if quals != None:
|
|
716 quals = quals[ :-count ]
|
|
717 count, op = cigar[ -2 ]
|
|
718 # Soft clip suffix
|
|
719 if op == "S":
|
|
720 nucs = nucs[ :-count ] + nucs[ -count: ].lower()
|
|
721 return nucs, quals
|
|
722
|
|
723 def samify_phred_scores( quals ):
|
|
724 """
|
|
725 Convert a decimal list of phred base-quality scores to a sam quality string.
|
|
726 Note that if a quality is outside the dynamic range of sam's ability to
|
|
727 represent it, we clip the value to the max allowed. SAM quality scores
|
|
728 range from chr(33) to chr(126).
|
|
729 """
|
|
730 if min( quals ) >= 0 and max( quals ) <= 126-33:
|
|
731 return "".join( [ chr( 33 + q ) for q in quals ] )
|
|
732 else:
|
|
733 return "".join( [ chr( max( 33, min( 126, 33+q ) ) ) for q in quals ] )
|
|
734
|
|
735 def reverse_complement( nucs ):
|
|
736 complementMap = maketrans( "ACGTacgt", "TGCAtgca" )
|
|
737 return nucs[ ::-1 ].translate( complementMap )
|
|
738
|
|
739 def reverse_string( s ):
|
|
740 return s[ ::-1 ]
|
|
741
|
|
742 def __main__():
|
|
743 # Parse command line
|
|
744 # input1: a reference genome ( 2bit or fasta )
|
|
745 # input2: a collection of 454 paired end reads ( a fasta file )
|
|
746 # input3: a linker sequence ( a very small fasta file )
|
|
747 # input4: a base quality score 454 file ( qual454 )
|
|
748 parser = optparse.OptionParser()
|
|
749 parser.add_option( '', '--ref_name', dest='ref_name', help='The reference name to change all output matches to' )
|
|
750 parser.add_option( '', '--ref_source', dest='ref_source', help='The reference is cached or from the history' )
|
|
751 parser.add_option( '', '--ref_sequences', dest='ref_sequences', help='Number of sequences in the reference dataset' )
|
|
752 parser.add_option( '', '--source_select', dest='source_select', help='Use pre-set or cached reference file' )
|
|
753 parser.add_option( '', '--input1', dest='input1', help='The name of the reference file if using history or reference base name if using cached' )
|
|
754 parser.add_option( '', '--input2', dest='input2', help='The 454 reads file to align' )
|
|
755 parser.add_option( '', '--input3', dest='input3', help='The sequencing linker file' )
|
|
756 parser.add_option( '', '--input4', dest='input4', help='The base quality score 454 file' )
|
|
757 parser.add_option( '', '--output', dest='output', help='The output file' )
|
|
758 parser.add_option( '', '--lastz_seqs_file_dir', dest='lastz_seqs_file_dir', help='Directory of local lastz_seqs.loc file' )
|
|
759 ( options, args ) = parser.parse_args()
|
|
760
|
|
761 # output version # of tool
|
|
762 try:
|
|
763 tmp = tempfile.NamedTemporaryFile().name
|
|
764 tmp_stdout = open( tmp, 'wb' )
|
|
765 proc = subprocess.Popen( args='lastz -v', shell=True, stdout=tmp_stdout )
|
|
766 tmp_stdout.close()
|
|
767 returncode = proc.wait()
|
|
768 stdout = None
|
|
769 for line in open( tmp_stdout.name, 'rb' ):
|
|
770 if line.lower().find( 'version' ) >= 0:
|
|
771 stdout = line.strip()
|
|
772 break
|
|
773 if stdout:
|
|
774 sys.stdout.write( '%s\n' % stdout )
|
|
775 else:
|
|
776 raise Exception
|
|
777 except:
|
|
778 sys.stdout.write( 'Could not determine Lastz version\n' )
|
|
779
|
|
780 if options.ref_name:
|
|
781 ref_name = '[nickname=%s]' % options.ref_name
|
|
782 else:
|
|
783 ref_name = ''
|
|
784 if options.ref_source == 'history':
|
|
785 # Reference is a fasta dataset from the history
|
|
786 try:
|
|
787 # Ensure there is at least 1 sequence in the dataset ( this may not be necessary ).
|
|
788 error_msg = "The reference dataset is missing metadata, click the pencil icon in the history item and 'auto-detect' the metadata attributes."
|
|
789 ref_sequences = int( options.ref_sequences )
|
|
790 if ref_sequences < 1:
|
|
791 stop_err( error_msg )
|
|
792 except:
|
|
793 stop_err( error_msg )
|
|
794 else:
|
|
795 ref_sequences = 0
|
|
796 tmp_w12_name = get_tmp_file_name( suffix='vs_linker.W12' )
|
|
797 tmp_T1_name = get_tmp_file_name( suffix='vs_linker.T1' )
|
|
798 # Run lastz twice ( with different options ) on the linker sequence and paired end reads,
|
|
799 # looking for the linker ( each run finds some the other doesn't )
|
|
800 command = 'lastz %s %s W=12 --notrans --exact=18 --match=1,3 O=1 E=3 Y=10 L=18 --ambiguousn --coverage=85 --format=general-:name2,zstart2+,length2,size2 > %s' % \
|
|
801 ( options.input3, options.input2, tmp_w12_name )
|
|
802 run_command( command )
|
|
803 command = 'lastz %s %s T=1 --match=1,2 O=1 E=2 X=15 K=10 Y=15 L=18 --ambiguousn --coverage=85 --format=general-:name2,zstart2+,length2,size2 > %s' % \
|
|
804 ( options.input3, options.input2, tmp_T1_name )
|
|
805 run_command( command )
|
|
806 # Combine the alignment output from the two lastz runs
|
|
807 tmp_combined_linker_file_name = get_tmp_file_name( suffix='vs_linker' )
|
|
808 command = 'cat %s %s | sort -u > %s' % ( tmp_w12_name, tmp_T1_name, tmp_combined_linker_file_name )
|
|
809 run_command( command )
|
|
810 # Use the alignment info to split reads into left and right mates
|
|
811 tmp_mates_mapping_file_name, tmp_mates_file_name, tmp_mates_short_file_name, tmp_mates_long_file_name = split_paired_reads( options.input2, tmp_combined_linker_file_name )
|
|
812 # Align mates to the reference - tmp_align_file_names is a list of file names created by align_mates()
|
|
813 tmp_align_file_name_list = align_mates( options.input1, options.ref_source, ref_name, ref_sequences, tmp_mates_short_file_name, tmp_mates_long_file_name )
|
|
814 # Combine and convert mate coordinates back to read coordinates
|
|
815 paired_mate_unmapper( options.input2, options.input4, tmp_mates_mapping_file_name, tmp_align_file_name_list, options.output )
|
|
816 # Delete all temporary files
|
|
817 for file_name in tmp_file_names:
|
|
818 os.remove( file_name )
|
|
819 # Handle any invalid lines in the input data
|
|
820 if total_skipped_lines:
|
|
821 msgs = dict( bad_interval="Bad interval in line",
|
|
822 inconsistent_read_lengths="Inconsistent read/quality lengths for seq #",
|
|
823 inconsistent_reads="Inconsistent reads for seq #",
|
|
824 inconsistent_sizes="Inconsistent sizes for seq #",
|
|
825 missing_mate="Mapping file does not include mate on line",
|
|
826 missing_quals="Missing quality values for name on line",
|
|
827 missing_seq="Missing sequence for name on line",
|
|
828 multiple_seqs="Multiple names for seq #",
|
|
829 no_header="First quality sequence has no header",
|
|
830 num_fields="Must have 4 fields in line",
|
|
831 reads_paired="SAM flag indicates reads already paired on line",
|
|
832 sam_flag="Bad SAM flag on line",
|
|
833 sam_headers="SAM headers on line",
|
|
834 sam_min_columns="Need 11 columns on line",
|
|
835 two_mate_names="Mate name already seen, line",
|
|
836 wrong_seq_len="Size differs from length of seq #" )
|
|
837 print "Skipped %d invalid lines: "
|
|
838 msg = ""
|
|
839 for k, v in skipped_lines.items():
|
|
840 if v[0]:
|
|
841 # v[0] is the number of times the error occurred
|
|
842 # v[1] is the position of the line or sequence in the file
|
|
843 # v[2] is the name of the sequence or the text of the line
|
|
844 msg += "(%d)%s %d:%s. " % ( v[0], msgs[k], v[1], v[2] )
|
|
845 print msg
|
|
846
|
|
847 if __name__=="__main__": __main__()
|