renamed to hopefully clearer express functionality
This commit is contained in:
parent
0f04ddd88e
commit
814b05e2f6
|
@ -14,18 +14,19 @@ scriptID = ' '.join([scriptName,damask.version])
|
||||||
# --------------------------------------------------------------------
|
# --------------------------------------------------------------------
|
||||||
|
|
||||||
parser = OptionParser(option_class=damask.extendableOption, usage='%prog options [file[s]]', description = """
|
parser = OptionParser(option_class=damask.extendableOption, usage='%prog options [file[s]]', description = """
|
||||||
Add data in column(s) of second ASCIItable selected from the row indexed by the value in a mapping column.
|
Add data in column(s) of mapped ASCIItable selected from the row indexed by the value in a mapping column.
|
||||||
|
Row numbers start at 1.
|
||||||
|
|
||||||
""", version = scriptID)
|
""", version = scriptID)
|
||||||
|
|
||||||
parser.add_option('-c','--map',
|
parser.add_option('--index',
|
||||||
dest = 'map',
|
dest = 'index',
|
||||||
type = 'string', metavar = 'string',
|
type = 'string', metavar = 'string',
|
||||||
help = 'column label containing row mapping')
|
help = 'column label containing row index')
|
||||||
parser.add_option('-o','--offset',
|
parser.add_option('-o','--offset',
|
||||||
dest = 'offset',
|
dest = 'offset',
|
||||||
type = 'int', metavar = 'int',
|
type = 'int', metavar = 'int',
|
||||||
help = 'offset between mapping column value and actual row in mapped table [%default]')
|
help = 'constant offset for index column value [%default]')
|
||||||
parser.add_option('-l','--label',
|
parser.add_option('-l','--label',
|
||||||
dest = 'label',
|
dest = 'label',
|
||||||
action = 'extend', metavar = '<string LIST>',
|
action = 'extend', metavar = '<string LIST>',
|
||||||
|
@ -42,19 +43,19 @@ parser.set_defaults(offset = 0,
|
||||||
|
|
||||||
if options.label is None:
|
if options.label is None:
|
||||||
parser.error('no data columns specified.')
|
parser.error('no data columns specified.')
|
||||||
if options.map is None:
|
if options.index is None:
|
||||||
parser.error('no mapping column given.')
|
parser.error('no index column given.')
|
||||||
|
|
||||||
# ------------------------------------------ process mapping ASCIItable ---------------------------
|
# ------------------------------------------ process indexed ASCIItable ---------------------------
|
||||||
|
|
||||||
if options.asciitable is not None and os.path.isfile(options.asciitable):
|
if options.asciitable is not None and os.path.isfile(options.asciitable):
|
||||||
|
|
||||||
mappedTable = damask.ASCIItable(name = options.asciitable,
|
indexedTable = damask.ASCIItable(name = options.asciitable,
|
||||||
buffered = False,
|
buffered = False,
|
||||||
readonly = True)
|
readonly = True)
|
||||||
mappedTable.head_read() # read ASCII header info of mapped table
|
indexedTable.head_read() # read ASCII header info of indexed table
|
||||||
missing_labels = mappedTable.data_readArray(options.label)
|
missing_labels = indexedTable.data_readArray(options.label)
|
||||||
mappedTable.close() # close mapped input ASCII table
|
indexedTable.close() # close input ASCII table
|
||||||
|
|
||||||
if len(missing_labels) > 0:
|
if len(missing_labels) > 0:
|
||||||
damask.util.croak('column{} {} not found...'.format('s' if len(missing_labels) > 1 else '',', '.join(missing_labels)))
|
damask.util.croak('column{} {} not found...'.format('s' if len(missing_labels) > 1 else '',', '.join(missing_labels)))
|
||||||
|
@ -80,8 +81,8 @@ for name in filenames:
|
||||||
|
|
||||||
errors = []
|
errors = []
|
||||||
|
|
||||||
mappedColumn = table.label_index(options.map)
|
indexColumn = table.label_index(options.index)
|
||||||
if mappedColumn < 0: errors.append('mapping column {} not found.'.format(options.map))
|
if indexColumn < 0: errors.append('index column {} not found.'.format(options.index))
|
||||||
|
|
||||||
if errors != []:
|
if errors != []:
|
||||||
damask.util.croak(errors)
|
damask.util.croak(errors)
|
||||||
|
@ -91,7 +92,7 @@ for name in filenames:
|
||||||
# ------------------------------------------ assemble header --------------------------------------
|
# ------------------------------------------ assemble header --------------------------------------
|
||||||
|
|
||||||
table.info_append(scriptID + '\t' + ' '.join(sys.argv[1:]))
|
table.info_append(scriptID + '\t' + ' '.join(sys.argv[1:]))
|
||||||
table.labels_append(mappedTable.labels(raw = True)) # extend ASCII header with new labels
|
table.labels_append(indexedTable.labels(raw = True)) # extend ASCII header with new labels
|
||||||
table.head_write()
|
table.head_write()
|
||||||
|
|
||||||
# ------------------------------------------ process data ------------------------------------------
|
# ------------------------------------------ process data ------------------------------------------
|
||||||
|
@ -99,9 +100,9 @@ for name in filenames:
|
||||||
outputAlive = True
|
outputAlive = True
|
||||||
while outputAlive and table.data_read(): # read next data line of ASCII table
|
while outputAlive and table.data_read(): # read next data line of ASCII table
|
||||||
try:
|
try:
|
||||||
table.data_append(mappedTable.data[int(round(float(table.data[mappedColumn])))+options.offset-1]) # add all mapped data types
|
table.data_append(indexedTable.data[int(round(float(table.data[indexColumn])))+options.offset-1]) # add all mapped data types
|
||||||
except IndexError:
|
except IndexError:
|
||||||
table.data_append(np.nan*np.ones_like(mappedTable.data[0]))
|
table.data_append(np.nan*np.ones_like(indexedTable.data[0]))
|
||||||
outputAlive = table.data_write() # output processed line
|
outputAlive = table.data_write() # output processed line
|
||||||
|
|
||||||
# ------------------------------------------ output finalization -----------------------------------
|
# ------------------------------------------ output finalization -----------------------------------
|
|
@ -14,12 +14,12 @@ scriptID = ' '.join([scriptName,damask.version])
|
||||||
# --------------------------------------------------------------------
|
# --------------------------------------------------------------------
|
||||||
|
|
||||||
parser = OptionParser(option_class=damask.extendableOption, usage='%prog options [file[s]]', description = """
|
parser = OptionParser(option_class=damask.extendableOption, usage='%prog options [file[s]]', description = """
|
||||||
Add data of selected column(s) from (first) row of second ASCIItable that shares the mapping column value.
|
Add data of selected column(s) from (first) row of second ASCIItable that shares the linking column value.
|
||||||
|
|
||||||
""", version = scriptID)
|
""", version = scriptID)
|
||||||
|
|
||||||
parser.add_option('-c','--map',
|
parser.add_option('--link',
|
||||||
dest = 'map', nargs = 2,
|
dest = 'link', nargs = 2,
|
||||||
type = 'string', metavar = 'string string',
|
type = 'string', metavar = 'string string',
|
||||||
help = 'column labels containing linked values')
|
help = 'column labels containing linked values')
|
||||||
parser.add_option('-l','--label',
|
parser.add_option('-l','--label',
|
||||||
|
@ -29,7 +29,7 @@ parser.add_option('-l','--label',
|
||||||
parser.add_option('-a','--asciitable',
|
parser.add_option('-a','--asciitable',
|
||||||
dest = 'asciitable',
|
dest = 'asciitable',
|
||||||
type = 'string', metavar = 'string',
|
type = 'string', metavar = 'string',
|
||||||
help = 'indexed ASCIItable')
|
help = 'linked ASCIItable')
|
||||||
|
|
||||||
parser.set_defaults()
|
parser.set_defaults()
|
||||||
|
|
||||||
|
@ -37,32 +37,32 @@ parser.set_defaults()
|
||||||
|
|
||||||
if options.label is None:
|
if options.label is None:
|
||||||
parser.error('no data columns specified.')
|
parser.error('no data columns specified.')
|
||||||
if options.map is None:
|
if options.link is None:
|
||||||
parser.error('no mapping columns given.')
|
parser.error('no linking columns given.')
|
||||||
|
|
||||||
# ------------------------------------------ process mapping ASCIItable ---------------------------
|
# -------------------------------------- process linked ASCIItable --------------------------------
|
||||||
|
|
||||||
if options.asciitable is not None and os.path.isfile(options.asciitable):
|
if options.asciitable is not None and os.path.isfile(options.asciitable):
|
||||||
|
|
||||||
mappedTable = damask.ASCIItable(name = options.asciitable,
|
linkedTable = damask.ASCIItable(name = options.asciitable,
|
||||||
buffered = False,
|
buffered = False,
|
||||||
readonly = True)
|
readonly = True)
|
||||||
mappedTable.head_read() # read ASCII header info of mapped table
|
linkedTable.head_read() # read ASCII header info of linked table
|
||||||
if mappedTable.label_dimension(options.map[1]) != 1:
|
if linkedTable.label_dimension(options.link[1]) != 1:
|
||||||
parser.error('mapping column {} needs to be scalar valued.'.format(options.map[1]))
|
parser.error('linking column {} needs to be scalar valued.'.format(options.link[1]))
|
||||||
|
|
||||||
missing_labels = mappedTable.data_readArray([options.map[1]]+options.label)
|
missing_labels = linkedTable.data_readArray([options.link[1]]+options.label)
|
||||||
mappedTable.close() # close mapped input ASCII table
|
linkedTable.close() # close linked ASCII table
|
||||||
|
|
||||||
if len(missing_labels) > 0:
|
if len(missing_labels) > 0:
|
||||||
damask.util.croak('column{} {} not found...'.format('s' if len(missing_labels) > 1 else '',', '.join(missing_labels)))
|
damask.util.croak('column{} {} not found...'.format('s' if len(missing_labels) > 1 else '',', '.join(missing_labels)))
|
||||||
|
|
||||||
index = mappedTable.data[:,0]
|
index = linkedTable.data[:,0]
|
||||||
data = mappedTable.data[:,1:]
|
data = linkedTable.data[:,1:]
|
||||||
else:
|
else:
|
||||||
parser.error('no indexed ASCIItable given.')
|
parser.error('no linked ASCIItable given.')
|
||||||
|
|
||||||
# --- loop over input files -------------------------------------------------------------------------
|
# --- loop over input files -----------------------------------------------------------------------
|
||||||
|
|
||||||
if filenames == []: filenames = [None]
|
if filenames == []: filenames = [None]
|
||||||
|
|
||||||
|
@ -80,8 +80,8 @@ for name in filenames:
|
||||||
|
|
||||||
errors = []
|
errors = []
|
||||||
|
|
||||||
mappedColumn = table.label_index(options.map[0])
|
linkColumn = table.label_index(options.link[0])
|
||||||
if mappedColumn < 0: errors.append('mapping column {} not found.'.format(options.map[0]))
|
if linkColumn < 0: errors.append('linking column {} not found.'.format(options.link[0]))
|
||||||
|
|
||||||
if errors != []:
|
if errors != []:
|
||||||
damask.util.croak(errors)
|
damask.util.croak(errors)
|
||||||
|
@ -91,7 +91,7 @@ for name in filenames:
|
||||||
# ------------------------------------------ assemble header --------------------------------------
|
# ------------------------------------------ assemble header --------------------------------------
|
||||||
|
|
||||||
table.info_append(scriptID + '\t' + ' '.join(sys.argv[1:]))
|
table.info_append(scriptID + '\t' + ' '.join(sys.argv[1:]))
|
||||||
table.labels_append(mappedTable.labels(raw = True)[1:]) # extend with new labels (except for mapped column)
|
table.labels_append(linkedTable.labels(raw = True)[1:]) # extend with new labels (except for linked column)
|
||||||
|
|
||||||
table.head_write()
|
table.head_write()
|
||||||
|
|
||||||
|
@ -100,7 +100,7 @@ for name in filenames:
|
||||||
outputAlive = True
|
outputAlive = True
|
||||||
while outputAlive and table.data_read(): # read next data line of ASCII table
|
while outputAlive and table.data_read(): # read next data line of ASCII table
|
||||||
try:
|
try:
|
||||||
table.data_append(data[np.argwhere(index == float(table.data[mappedColumn]))[0]]) # add data from first matching line
|
table.data_append(data[np.argwhere(index == float(table.data[linkColumn]))[0]]) # add data from first matching line
|
||||||
except IndexError:
|
except IndexError:
|
||||||
table.data_append(np.nan*np.ones_like(data[0])) # or add NaNs
|
table.data_append(np.nan*np.ones_like(data[0])) # or add NaNs
|
||||||
outputAlive = table.data_write() # output processed line
|
outputAlive = table.data_write() # output processed line
|
Loading…
Reference in New Issue