From 9821654aae448686d1fe43785839b546b7a9b02c Mon Sep 17 00:00:00 2001 From: Martin Diehl Date: Thu, 19 Jul 2018 16:16:10 +0200 Subject: [PATCH] iteritems does not exist in python3 anymore items also works for python2 (might be slower though) --- processing/post/addDeterminant.py | 4 ++-- processing/post/addDeviator.py | 6 +++--- processing/post/addGaussian.py | 6 +++--- processing/post/addMises.py | 4 ++-- processing/post/addSpectralDecomposition.py | 4 ++-- processing/post/addStrainTensors.py | 2 +- processing/post/rotateData.py | 2 +- processing/pre/geom_translate.py | 2 +- processing/pre/patchFromReconstructedBoundaries.py | 2 +- 9 files changed, 16 insertions(+), 16 deletions(-) diff --git a/processing/post/addDeterminant.py b/processing/post/addDeterminant.py index 1f721c27e..7196051e5 100755 --- a/processing/post/addDeterminant.py +++ b/processing/post/addDeterminant.py @@ -58,7 +58,7 @@ for name in filenames: errors = [] remarks = [] - for type, data in items.iteritems(): + for type, data in items.items(): for what in data['labels']: dim = table.label_dimension(what) if dim != data['dim']: remarks.append('column {} is not a {}...'.format(what,type)) @@ -81,7 +81,7 @@ for name in filenames: outputAlive = True while outputAlive and table.data_read(): # read next data line of ASCII table - for type, data in items.iteritems(): + for type, data in items.items(): for column in data['column']: table.data_append(determinant(map(float,table.data[column: column+data['dim']]))) diff --git a/processing/post/addDeviator.py b/processing/post/addDeviator.py index 471c2635f..4df8a6803 100755 --- a/processing/post/addDeviator.py +++ b/processing/post/addDeviator.py @@ -66,7 +66,7 @@ for name in filenames: remarks = [] column = {} - for type, data in items.iteritems(): + for type, data in items.items(): for what in data['labels']: dim = table.label_dimension(what) if dim != data['dim']: remarks.append('column {} is not a {}.'.format(what,type)) @@ -83,7 +83,7 @@ for name in filenames: # ------------------------------------------ assemble header -------------------------------------- table.info_append(scriptID + '\t' + ' '.join(sys.argv[1:])) - for type, data in items.iteritems(): + for type, data in items.items(): for label in data['active']: table.labels_append(['{}_dev({})'.format(i+1,label) for i in range(data['dim'])] + \ (['sph({})'.format(label)] if options.spherical else [])) # extend ASCII header with new labels @@ -93,7 +93,7 @@ for name in filenames: outputAlive = True while outputAlive and table.data_read(): # read next data line of ASCII table - for type, data in items.iteritems(): + for type, data in items.items(): for column in data['column']: table.data_append(deviator(map(float,table.data[column: column+data['dim']]),options.spherical)) diff --git a/processing/post/addGaussian.py b/processing/post/addGaussian.py index c198ef62f..bc0100f56 100755 --- a/processing/post/addGaussian.py +++ b/processing/post/addGaussian.py @@ -83,7 +83,7 @@ for name in filenames: if table.label_dimension(options.pos) != 3: errors.append('coordinates {} are not a vector.'.format(options.pos)) else: colCoord = table.label_index(options.pos) - for type, data in items.iteritems(): + for type, data in items.items(): for what in (data['labels'] if data['labels'] is not None else []): dim = table.label_dimension(what) if dim != data['dim']: remarks.append('column {} is not a {}.'.format(what,type)) @@ -100,7 +100,7 @@ for name in filenames: # ------------------------------------------ assemble header -------------------------------------- table.info_append(scriptID + '\t' + ' '.join(sys.argv[1:])) - for type, data in items.iteritems(): + for type, data in items.items(): for label in data['active']: table.labels_append(['Gauss{}({})'.format(options.sigma,label)]) # extend ASCII header with new labels table.head_write() @@ -114,7 +114,7 @@ for name in filenames: # ------------------------------------------ process value field ----------------------------------- stack = [table.data] - for type, data in items.iteritems(): + for type, data in items.items(): for i,label in enumerate(data['active']): stack.append(ndimage.filters.gaussian_filter(table.data[:,data['column'][i]], options.sigma,options.order, diff --git a/processing/post/addMises.py b/processing/post/addMises.py index 2ce350dbd..4719c2e35 100755 --- a/processing/post/addMises.py +++ b/processing/post/addMises.py @@ -70,7 +70,7 @@ for name in filenames: errors = [] remarks = [] - for type, data in items.iteritems(): + for type, data in items.items(): for what in data['labels']: dim = table.label_dimension(what) if dim != data['dim']: remarks.append('column {} is not a {}...'.format(what,type)) @@ -94,7 +94,7 @@ for name in filenames: outputAlive = True while outputAlive and table.data_read(): # read next data line of ASCII table - for type, data in items.iteritems(): + for type, data in items.items(): for column in data['column']: table.data_append(Mises(type, np.array(table.data[column:column+data['dim']],'d').reshape(data['shape']))) diff --git a/processing/post/addSpectralDecomposition.py b/processing/post/addSpectralDecomposition.py index 76bf2e875..b21900c0c 100755 --- a/processing/post/addSpectralDecomposition.py +++ b/processing/post/addSpectralDecomposition.py @@ -58,7 +58,7 @@ for name in filenames: errors = [] remarks = [] - for type, data in items.iteritems(): + for type, data in items.items(): for what in data['labels']: dim = table.label_dimension(what) if dim != data['dim']: remarks.append('column {} is not a {}...'.format(what,type)) @@ -84,7 +84,7 @@ for name in filenames: outputAlive = True while outputAlive and table.data_read(): # read next data line of ASCII table - for type, data in items.iteritems(): + for type, data in items.items(): for column in data['column']: (u,v) = np.linalg.eigh(np.array(map(float,table.data[column:column+data['dim']])).reshape(data['shape'])) if options.rh and np.dot(np.cross(v[:,0], v[:,1]), v[:,2]) < 0.0 : v[:, 2] *= -1.0 # ensure right-handed eigenvector basis diff --git a/processing/post/addStrainTensors.py b/processing/post/addStrainTensors.py index 447ae03ba..7cb9f3079 100755 --- a/processing/post/addStrainTensors.py +++ b/processing/post/addStrainTensors.py @@ -101,7 +101,7 @@ for name in filenames: errors = [] remarks = [] - for type, data in items.iteritems(): + for type, data in items.items(): for what in data['labels']: dim = table.label_dimension(what) if dim != data['dim']: remarks.append('column {} is not a {}...'.format(what,type)) diff --git a/processing/post/rotateData.py b/processing/post/rotateData.py index 08958cc86..c5e4882b9 100755 --- a/processing/post/rotateData.py +++ b/processing/post/rotateData.py @@ -73,7 +73,7 @@ for name in filenames: remarks = [] column = {} - for type, data in items.iteritems(): + for type, data in items.items(): for what in data['labels']: dim = table.label_dimension(what) if dim != data['dim']: remarks.append('column {} is not a {}.'.format(what,type)) diff --git a/processing/pre/geom_translate.py b/processing/pre/geom_translate.py index f8f6e4169..2f4918632 100755 --- a/processing/pre/geom_translate.py +++ b/processing/pre/geom_translate.py @@ -92,7 +92,7 @@ for name in filenames: } substituted = np.copy(microstructure) - for k, v in sub.iteritems(): substituted[microstructure==k] = v # substitute microstructure indices + for k, v in sub.items(): substituted[microstructure==k] = v # substitute microstructure indices substituted += options.microstructure # shift microstructure indices diff --git a/processing/pre/patchFromReconstructedBoundaries.py b/processing/pre/patchFromReconstructedBoundaries.py index a43ccc236..fabec0fdf 100755 --- a/processing/pre/patchFromReconstructedBoundaries.py +++ b/processing/pre/patchFromReconstructedBoundaries.py @@ -344,7 +344,7 @@ def rcbParser(content,M,size,tolerance,idcolumn,segmentcolumn): else: myNeighbors[grainNeighbors[leg][side]] = 1 if myNeighbors: # do I have any neighbors (i.e., non-bounding box segment) - candidateGrains = sorted(myNeighbors.iteritems(), key=lambda p: (p[1],p[0]), reverse=True) # sort grain counting + candidateGrains = sorted(myNeighbors.items(), key=lambda p: (p[1],p[0]), reverse=True) # sort grain counting # most frequent one not yet seen? rcData['grainMapping'].append(candidateGrains[0 if candidateGrains[0][0] not in rcData['grainMapping'] else 1][0]) # must be me then # special case of bi-crystal situation...