default format for Table is ASCII
hence, renamed load_ASCII/save_ASCII to load/save
This commit is contained in:
parent
5895e74029
commit
6ab88aad2b
2
PRIVATE
2
PRIVATE
|
@ -1 +1 @@
|
|||
Subproject commit 576f3af61c3d893b608809cc91e46647809010f1
|
||||
Subproject commit 7c543a98e89840f1f1540c7af8c62a19084dab6e
|
|
@ -42,10 +42,10 @@ rot_to_TSL = damask.Rotation.from_axis_angle([-1,0,0,.75*np.pi])
|
|||
for name in filenames:
|
||||
damask.util.report(scriptName,name)
|
||||
|
||||
table = damask.Table.load_ASCII(StringIO(''.join(sys.stdin.read())) if name is None else name)
|
||||
table = damask.Table.load(StringIO(''.join(sys.stdin.read())) if name is None else name)
|
||||
|
||||
coord = - table.get(options.frame)
|
||||
coord[:,2] += table.get(options.depth)[:,0]
|
||||
|
||||
table.add('coord',rot_to_TSL.broadcast_to(coord.shape[0]) @ coord,scriptID+' '+' '.join(sys.argv[1:]))\
|
||||
.save_ASCII((sys.stdout if name is None else name),legacy=True)
|
||||
.save((sys.stdout if name is None else name),legacy=True)
|
||||
|
|
|
@ -39,10 +39,10 @@ if options.labels is None:
|
|||
for name in filenames:
|
||||
damask.util.report(scriptName,name)
|
||||
|
||||
table = damask.Table.load_ASCII(StringIO(''.join(sys.stdin.read())) if name is None else name)
|
||||
table = damask.Table.load(StringIO(''.join(sys.stdin.read())) if name is None else name)
|
||||
for label in options.labels:
|
||||
table = table.add('cum_{}({})'.format('prod' if options.product else 'sum',label),
|
||||
np.cumprod(table.get(label),0) if options.product else np.cumsum(table.get(label),0),
|
||||
scriptID+' '+' '.join(sys.argv[1:]))
|
||||
|
||||
table.save_ASCII((sys.stdout if name is None else name),legacy=True)
|
||||
table.save((sys.stdout if name is None else name),legacy=True)
|
||||
|
|
|
@ -60,4 +60,4 @@ for filename in options.filenames:
|
|||
os.mkdir(dirname,0o755)
|
||||
file_out = '{}_inc{}.txt'.format(os.path.splitext(os.path.split(filename)[-1])[0],
|
||||
inc[3:].zfill(N_digits))
|
||||
table.save_ASCII(os.path.join(dirname,file_out),legacy=True)
|
||||
table.save(os.path.join(dirname,file_out),legacy=True)
|
||||
|
|
|
@ -172,7 +172,7 @@ if filenames == []: filenames = [None]
|
|||
for name in filenames:
|
||||
damask.util.report(scriptName,name)
|
||||
|
||||
table = damask.Table.load_ASCII(StringIO(''.join(sys.stdin.read())) if name is None else name)
|
||||
table = damask.Table.load(StringIO(''.join(sys.stdin.read())) if name is None else name)
|
||||
grid,size,origin = damask.grid_filters.cell_coord0_gridSizeOrigin(table.get(options.pos))
|
||||
|
||||
F = table.get(options.defgrad).reshape(tuple(grid)+(-1,),order='F').reshape(tuple(grid)+(3,3))
|
||||
|
@ -191,4 +191,4 @@ for name in filenames:
|
|||
volumeMismatch.reshape(-1,1,order='F'),
|
||||
scriptID+' '+' '.join(sys.argv[1:]))
|
||||
|
||||
table.save_ASCII((sys.stdout if name is None else name), legacy=True)
|
||||
table.save((sys.stdout if name is None else name), legacy=True)
|
||||
|
|
|
@ -43,7 +43,7 @@ if options.labels is None: parser.error('no data column specified.')
|
|||
for name in filenames:
|
||||
damask.util.report(scriptName,name)
|
||||
|
||||
table = damask.Table.load_ASCII(StringIO(''.join(sys.stdin.read())) if name is None else name)
|
||||
table = damask.Table.load(StringIO(''.join(sys.stdin.read())) if name is None else name)
|
||||
grid,size,origin = damask.grid_filters.cell_coord0_gridSizeOrigin(table.get(options.pos))
|
||||
|
||||
for label in options.labels:
|
||||
|
@ -55,4 +55,4 @@ for name in filenames:
|
|||
curl.reshape(tuple(grid)+(-1,)).reshape(-1,np.prod(shape),order='F'),
|
||||
scriptID+' '+' '.join(sys.argv[1:]))
|
||||
|
||||
table.save_ASCII((sys.stdout if name is None else name), legacy=True)
|
||||
table.save((sys.stdout if name is None else name), legacy=True)
|
||||
|
|
|
@ -65,10 +65,10 @@ if options.labels is None:
|
|||
for name in filenames:
|
||||
damask.util.report(scriptName,name)
|
||||
|
||||
table = damask.Table.load_ASCII(StringIO(''.join(sys.stdin.read())) if name is None else name)
|
||||
table = damask.Table.load(StringIO(''.join(sys.stdin.read())) if name is None else name)
|
||||
for label in options.labels:
|
||||
table = table.add('d({})/d({})'.format(label,options.coordinates),
|
||||
derivative(table.get(options.coordinates),table.get(label)),
|
||||
scriptID+' '+' '.join(sys.argv[1:]))
|
||||
|
||||
table.save_ASCII((sys.stdout if name is None else name), legacy=True)
|
||||
table.save((sys.stdout if name is None else name), legacy=True)
|
||||
|
|
|
@ -47,7 +47,7 @@ parser.set_defaults(f = 'f',
|
|||
for name in filenames:
|
||||
damask.util.report(scriptName,name)
|
||||
|
||||
table = damask.Table.load_ASCII(StringIO(''.join(sys.stdin.read())) if name is None else name)
|
||||
table = damask.Table.load(StringIO(''.join(sys.stdin.read())) if name is None else name)
|
||||
grid,size,origin = damask.grid_filters.cell_coord0_gridSizeOrigin(table.get(options.pos))
|
||||
|
||||
F = table.get(options.f).reshape(tuple(grid)+(-1,),order='F').reshape(tuple(grid)+(3,3))
|
||||
|
@ -60,7 +60,7 @@ for name in filenames:
|
|||
.add('fluct({}).{}'.format(options.f,options.pos),
|
||||
damask.grid_filters.node_displacement_fluct(size,F).reshape(-1,3,order='F'),
|
||||
scriptID+' '+' '.join(sys.argv[1:]))\
|
||||
.save_ASCII((sys.stdout if name is None else os.path.splitext(name)[0]+'_nodal.txt'), legacy=True)
|
||||
.save((sys.stdout if name is None else os.path.splitext(name)[0]+'_nodal.txt'), legacy=True)
|
||||
else:
|
||||
table.add('avg({}).{}'.format(options.f,options.pos),
|
||||
damask.grid_filters.cell_displacement_avg(size,F).reshape(-1,3,order='F'),
|
||||
|
@ -68,4 +68,4 @@ for name in filenames:
|
|||
.add('fluct({}).{}'.format(options.f,options.pos),
|
||||
damask.grid_filters.cell_displacement_fluct(size,F).reshape(-1,3,order='F'),
|
||||
scriptID+' '+' '.join(sys.argv[1:]))\
|
||||
.save_ASCII((sys.stdout if name is None else name), legacy=True)
|
||||
.save((sys.stdout if name is None else name), legacy=True)
|
||||
|
|
|
@ -43,7 +43,7 @@ if options.labels is None: parser.error('no data column specified.')
|
|||
for name in filenames:
|
||||
damask.util.report(scriptName,name)
|
||||
|
||||
table = damask.Table.load_ASCII(StringIO(''.join(sys.stdin.read())) if name is None else name)
|
||||
table = damask.Table.load(StringIO(''.join(sys.stdin.read())) if name is None else name)
|
||||
grid,size,origin = damask.grid_filters.cell_coord0_gridSizeOrigin(table.get(options.pos))
|
||||
|
||||
for label in options.labels:
|
||||
|
@ -55,4 +55,4 @@ for name in filenames:
|
|||
div.reshape(tuple(grid)+(-1,)).reshape(-1,np.prod(shape)//3,order='F'),
|
||||
scriptID+' '+' '.join(sys.argv[1:]))
|
||||
|
||||
table.save_ASCII((sys.stdout if name is None else name), legacy=True)
|
||||
table.save((sys.stdout if name is None else name), legacy=True)
|
||||
|
|
|
@ -142,7 +142,7 @@ for i,feature in enumerate(features):
|
|||
for name in filenames:
|
||||
damask.util.report(scriptName,name)
|
||||
|
||||
table = damask.Table.load_ASCII(StringIO(''.join(sys.stdin.read())) if name is None else name)
|
||||
table = damask.Table.load(StringIO(''.join(sys.stdin.read())) if name is None else name)
|
||||
grid,size,origin = damask.grid_filters.cell_coord0_gridSizeOrigin(table.get(options.pos))
|
||||
|
||||
neighborhood = neighborhoods[options.neighborhood]
|
||||
|
@ -184,4 +184,4 @@ for name in filenames:
|
|||
distance[i,:],
|
||||
scriptID+' '+' '.join(sys.argv[1:]))
|
||||
|
||||
table.save_ASCII((sys.stdout if name is None else name), legacy=True)
|
||||
table.save((sys.stdout if name is None else name), legacy=True)
|
||||
|
|
|
@ -63,7 +63,7 @@ if options.labels is None: parser.error('no data column specified.')
|
|||
for name in filenames:
|
||||
damask.util.report(scriptName,name)
|
||||
|
||||
table = damask.Table.load_ASCII(StringIO(''.join(sys.stdin.read())) if name is None else name)
|
||||
table = damask.Table.load(StringIO(''.join(sys.stdin.read())) if name is None else name)
|
||||
damask.grid_filters.coord0_check(table.get(options.pos))
|
||||
|
||||
for label in options.labels:
|
||||
|
@ -73,4 +73,4 @@ for name in filenames:
|
|||
mode = 'wrap' if options.periodic else 'nearest'),
|
||||
scriptID+' '+' '.join(sys.argv[1:]))
|
||||
|
||||
table.save_ASCII((sys.stdout if name is None else name), legacy=True)
|
||||
table.save((sys.stdout if name is None else name), legacy=True)
|
||||
|
|
|
@ -43,7 +43,7 @@ if options.labels is None: parser.error('no data column specified.')
|
|||
for name in filenames:
|
||||
damask.util.report(scriptName,name)
|
||||
|
||||
table = damask.Table.load_ASCII(StringIO(''.join(sys.stdin.read())) if name is None else name)
|
||||
table = damask.Table.load(StringIO(''.join(sys.stdin.read())) if name is None else name)
|
||||
grid,size,origin = damask.grid_filters.cell_coord0_gridSizeOrigin(table.get(options.pos))
|
||||
|
||||
for label in options.labels:
|
||||
|
@ -55,4 +55,4 @@ for name in filenames:
|
|||
grad.reshape(tuple(grid)+(-1,)).reshape(-1,np.prod(shape)*3,order='F'),
|
||||
scriptID+' '+' '.join(sys.argv[1:]))
|
||||
|
||||
table.save_ASCII((sys.stdout if name is None else name), legacy=True)
|
||||
table.save((sys.stdout if name is None else name), legacy=True)
|
||||
|
|
|
@ -110,7 +110,7 @@ R = damask.Rotation.from_axis_angle(np.array(options.labrotation),options.degree
|
|||
for name in filenames:
|
||||
damask.util.report(scriptName,name)
|
||||
|
||||
table = damask.Table.load_ASCII(StringIO(''.join(sys.stdin.read())) if name is None else name)
|
||||
table = damask.Table.load(StringIO(''.join(sys.stdin.read())) if name is None else name)
|
||||
|
||||
if options.eulers is not None:
|
||||
label = options.eulers
|
||||
|
@ -147,4 +147,4 @@ for name in filenames:
|
|||
if 'axisangle' in options.output:
|
||||
table = table.add('om({})'.format(label),o.as_axisangle(options.degrees), scriptID+' '+' '.join(sys.argv[1:]))
|
||||
|
||||
table.save_ASCII((sys.stdout if name is None else name), legacy=True)
|
||||
table.save((sys.stdout if name is None else name), legacy=True)
|
||||
|
|
|
@ -175,7 +175,7 @@ labels = ['S[{direction[0]:.1g}_{direction[1]:.1g}_{direction[2]:.1g}]'
|
|||
for name in filenames:
|
||||
damask.util.report(scriptName,name)
|
||||
|
||||
table = damask.Table.load_ASCII(StringIO(''.join(sys.stdin.read())) if name is None else name)
|
||||
table = damask.Table.load(StringIO(''.join(sys.stdin.read())) if name is None else name)
|
||||
|
||||
o = damask.Rotation.from_quaternion(table.get(options.quaternion))
|
||||
|
||||
|
@ -189,4 +189,4 @@ for name in filenames:
|
|||
for i,label in enumerate(labels):
|
||||
table = table.add(label,S[:,i],scriptID+' '+' '.join(sys.argv[1:]))
|
||||
|
||||
table.save_ASCII((sys.stdout if name is None else name), legacy=True)
|
||||
table.save((sys.stdout if name is None else name), legacy=True)
|
||||
|
|
|
@ -47,7 +47,7 @@ if filenames == []: filenames = [None]
|
|||
for name in filenames:
|
||||
damask.util.report(scriptName,name)
|
||||
|
||||
table = damask.Table.load_ASCII(StringIO(''.join(sys.stdin.read())) if name is None else name)
|
||||
table = damask.Table.load(StringIO(''.join(sys.stdin.read())) if name is None else name)
|
||||
|
||||
randomSeed = int(os.urandom(4).hex(), 16) if options.randomSeed is None else options.randomSeed # random seed per file
|
||||
rng = np.random.default_rng(randomSeed)
|
||||
|
@ -58,4 +58,4 @@ for name in filenames:
|
|||
rng.shuffle(uniques)
|
||||
table = table.set(label,uniques[inverse], scriptID+' '+' '.join(sys.argv[1:]))
|
||||
|
||||
table.save_ASCII((sys.stdout if name is None else name), legacy=True)
|
||||
table.save((sys.stdout if name is None else name), legacy=True)
|
||||
|
|
|
@ -154,4 +154,4 @@ for name in filenames:
|
|||
homogenization=options.homogenization,comments=header)
|
||||
damask.util.croak(geom)
|
||||
|
||||
geom.save_ASCII(os.path.splitext(name)[0]+'.geom',pack=False)
|
||||
geom.save_ASCII(os.path.splitext(name)[0]+'.geom',compress=False)
|
||||
|
|
|
@ -68,7 +68,7 @@ if options.axes is not None and not set(options.axes).issubset(set(['x','+x','-x
|
|||
for name in filenames:
|
||||
damask.util.report(scriptName,name)
|
||||
|
||||
table = damask.Table.load_ASCII(StringIO(''.join(sys.stdin.read())) if name is None else name)
|
||||
table = damask.Table.load(StringIO(''.join(sys.stdin.read())) if name is None else name)
|
||||
table.sort_by(['{}_{}'.format(i,options.pos) for i in range(3,0,-1)]) # x fast, y slow
|
||||
grid,size,origin = damask.grid_filters.cell_coord0_gridSizeOrigin(table.get(options.pos))
|
||||
|
||||
|
|
|
@ -171,7 +171,7 @@ if filenames == []: filenames = [None]
|
|||
for name in filenames:
|
||||
damask.util.report(scriptName,name)
|
||||
|
||||
table = damask.Table.load_ASCII(StringIO(''.join(sys.stdin.read())) if name is None else name)
|
||||
table = damask.Table.load(StringIO(''.join(sys.stdin.read())) if name is None else name)
|
||||
|
||||
size = np.ones(3)
|
||||
origin = np.zeros(3)
|
||||
|
@ -228,4 +228,4 @@ for name in filenames:
|
|||
homogenization=options.homogenization,comments=header)
|
||||
damask.util.croak(geom)
|
||||
|
||||
geom.save_ASCII(sys.stdout if name is None else os.path.splitext(name)[0]+'.geom',pack=False)
|
||||
geom.save_ASCII(sys.stdout if name is None else os.path.splitext(name)[0]+'.geom',compress=False)
|
||||
|
|
|
@ -172,4 +172,4 @@ for name in filenames:
|
|||
geom = geom.duplicate(microstructure[0:grid_original[0],0:grid_original[1],0:grid_original[2]])
|
||||
geom.add_comments(scriptID + ' ' + ' '.join(sys.argv[1:]))
|
||||
|
||||
geom.save_ASCII(sys.stdout if name is None else name,pack=False)
|
||||
geom.save_ASCII(sys.stdout if name is None else name,compress=False)
|
||||
|
|
|
@ -234,7 +234,7 @@ if filenames == []: filenames = [None]
|
|||
for name in filenames:
|
||||
damask.util.report(scriptName,name)
|
||||
|
||||
table = damask.Table.load_ASCII(StringIO(''.join(sys.stdin.read())) if name is None else name)
|
||||
table = damask.Table.load(StringIO(''.join(sys.stdin.read())) if name is None else name)
|
||||
|
||||
randomSeed = int(os.urandom(4).hex(),16) if options.randomSeed is None else options.randomSeed # random seed per file
|
||||
random.seed(randomSeed)
|
||||
|
|
|
@ -78,7 +78,7 @@ class myThread (threading.Thread):
|
|||
perturbedSeedsVFile = StringIO()
|
||||
myBestSeedsVFile.seek(0)
|
||||
|
||||
perturbedSeedsTable = damask.Table.load_ASCII(myBestSeedsVFile)
|
||||
perturbedSeedsTable = damask.Table.load(myBestSeedsVFile)
|
||||
coords = perturbedSeedsTable.get('pos')
|
||||
i = 0
|
||||
for ms,coord in enumerate(coords):
|
||||
|
@ -89,7 +89,7 @@ class myThread (threading.Thread):
|
|||
coords[i]=newCoords
|
||||
direction[i]*=2.
|
||||
i+= 1
|
||||
perturbedSeedsTable.set('pos',coords).save_ASCII(perturbedSeedsVFile,legacy=True)
|
||||
perturbedSeedsTable.set('pos',coords).save(perturbedSeedsVFile,legacy=True)
|
||||
|
||||
#--- do tesselation with perturbed seed file ------------------------------------------------------
|
||||
perturbedGeomVFile.close()
|
||||
|
|
|
@ -65,4 +65,4 @@ for name in filenames:
|
|||
|
||||
damask.Table(seeds[mask],{'pos':(3,)},comments)\
|
||||
.add('microstructure',microstructure[mask].astype(int))\
|
||||
.save_ASCII(sys.stdout if name is None else os.path.splitext(name)[0]+'.seeds',legacy=True)
|
||||
.save(sys.stdout if name is None else os.path.splitext(name)[0]+'.seeds',legacy=True)
|
||||
|
|
|
@ -92,5 +92,5 @@ for name in filenames:
|
|||
|
||||
table = damask.Table(seeds,{'pos':(3,),'microstructure':(1,)},comments)
|
||||
table.set('microstructure',table.get('microstructure').astype(np.int))\
|
||||
.save_ASCII(sys.stdout if name is None else \
|
||||
.save(sys.stdout if name is None else \
|
||||
os.path.splitext(name)[0]+f'_poked_{options.N}.seeds',legacy=True)
|
||||
|
|
|
@ -162,4 +162,4 @@ for name in filenames:
|
|||
else np.random.normal(loc = options.mean, scale = options.sigma, size = options.N)
|
||||
table = table.add('weight',weights)
|
||||
|
||||
table.save_ASCII(sys.stdout if name is None else name,legacy=True)
|
||||
table.save(sys.stdout if name is None else name,legacy=True)
|
||||
|
|
|
@ -297,9 +297,9 @@ class Colormap(mpl.colors.ListedColormap):
|
|||
|
||||
if fhandle is None:
|
||||
with open(self.name.replace(' ','_')+'.txt', 'w') as f:
|
||||
t.save_ASCII(f)
|
||||
t.save(f)
|
||||
else:
|
||||
t.save_ASCII(fhandle)
|
||||
t.save(fhandle)
|
||||
|
||||
|
||||
def save_GOM(self,fname=None):
|
||||
|
|
|
@ -67,7 +67,7 @@ class Table:
|
|||
|
||||
|
||||
@staticmethod
|
||||
def load_ASCII(fname):
|
||||
def load(fname):
|
||||
"""
|
||||
Load ASCII table file.
|
||||
|
||||
|
@ -344,7 +344,7 @@ class Table:
|
|||
return dup
|
||||
|
||||
|
||||
def save_ASCII(self,fname,legacy=False):
|
||||
def save(self,fname,legacy=False):
|
||||
"""
|
||||
Save as plain text file.
|
||||
|
||||
|
|
|
@ -287,9 +287,9 @@ class Test:
|
|||
|
||||
import numpy as np
|
||||
logging.info('\n '.join(['comparing',File1,File2]))
|
||||
table = damask.Table.load_ASCII(File1)
|
||||
table = damask.Table.load(File1)
|
||||
len1 = len(table.comments)+2
|
||||
table = damask.Table.load_ASCII(File2)
|
||||
table = damask.Table.load(File2)
|
||||
len2 = len(table.comments)+2
|
||||
|
||||
refArray = np.nan_to_num(np.genfromtxt(File1,missing_values='n/a',skip_header = len1,autostrip=True))
|
||||
|
@ -436,7 +436,7 @@ class Test:
|
|||
if not (isinstance(files, Iterable) and not isinstance(files, str)): # check whether list of files is requested
|
||||
files = [str(files)]
|
||||
|
||||
tables = [damask.Table.load_ASCII(filename) for filename in files]
|
||||
tables = [damask.Table.load(filename) for filename in files]
|
||||
for table in tables:
|
||||
table._label_discrete()
|
||||
|
||||
|
@ -486,7 +486,7 @@ class Test:
|
|||
|
||||
if len(files) < 2: return True # single table is always close to itself...
|
||||
|
||||
tables = [damask.Table.load_ASCII(filename) for filename in files]
|
||||
tables = [damask.Table.load(filename) for filename in files]
|
||||
|
||||
columns += [columns[0]]*(len(files)-len(columns)) # extend to same length as files
|
||||
columns = columns[:len(files)] # truncate to same length as files
|
||||
|
|
|
@ -106,8 +106,8 @@ class TestOrientation:
|
|||
coords = np.array([(1,i+1) for i,x in enumerate(eu)])
|
||||
table = Table(eu,{'Eulers':(3,)})
|
||||
table = table.add('pos',coords)
|
||||
table.save_ASCII(reference)
|
||||
assert np.allclose(eu,Table.load_ASCII(reference).get('Eulers'))
|
||||
table.save(reference)
|
||||
assert np.allclose(eu,Table.load(reference).get('Eulers'))
|
||||
|
||||
@pytest.mark.parametrize('lattice',Lattice.lattices)
|
||||
def test_disorientation360(self,lattice):
|
||||
|
|
|
@ -35,30 +35,30 @@ class TestTable:
|
|||
|
||||
@pytest.mark.parametrize('mode',['str','path'])
|
||||
def test_write_read(self,default,tmpdir,mode):
|
||||
default.save_ASCII(tmpdir/'default.txt')
|
||||
default.save(tmpdir/'default.txt')
|
||||
if mode == 'path':
|
||||
new = Table.load_ASCII(tmpdir/'default.txt')
|
||||
new = Table.load(tmpdir/'default.txt')
|
||||
elif mode == 'str':
|
||||
new = Table.load_ASCII(str(tmpdir/'default.txt'))
|
||||
new = Table.load(str(tmpdir/'default.txt'))
|
||||
assert all(default.data==new.data) and default.shapes == new.shapes
|
||||
|
||||
def test_write_read_file(self,default,tmpdir):
|
||||
with open(tmpdir/'default.txt','w') as f:
|
||||
default.save_ASCII(f)
|
||||
default.save(f)
|
||||
with open(tmpdir/'default.txt') as f:
|
||||
new = Table.load_ASCII(f)
|
||||
new = Table.load(f)
|
||||
assert all(default.data==new.data) and default.shapes == new.shapes
|
||||
|
||||
def test_write_read_legacy_style(self,default,tmpdir):
|
||||
with open(tmpdir/'legacy.txt','w') as f:
|
||||
default.save_ASCII(f,legacy=True)
|
||||
default.save(f,legacy=True)
|
||||
with open(tmpdir/'legacy.txt') as f:
|
||||
new = Table.load_ASCII(f)
|
||||
new = Table.load(f)
|
||||
assert all(default.data==new.data) and default.shapes == new.shapes
|
||||
|
||||
def test_write_invalid_format(self,default,tmpdir):
|
||||
with pytest.raises(TypeError):
|
||||
default.save_ASCII(tmpdir/'shouldnotbethere.txt',format='invalid')
|
||||
default.save(tmpdir/'shouldnotbethere.txt',format='invalid')
|
||||
|
||||
@pytest.mark.parametrize('mode',['str','path'])
|
||||
def test_read_ang(self,reference_dir,mode):
|
||||
|
@ -78,7 +78,7 @@ class TestTable:
|
|||
@pytest.mark.parametrize('fname',['datatype-mix.txt','whitespace-mix.txt'])
|
||||
def test_read_strange(self,reference_dir,fname):
|
||||
with open(reference_dir/fname) as f:
|
||||
Table.load_ASCII(f)
|
||||
Table.load(f)
|
||||
|
||||
def test_set(self,default):
|
||||
d = default.set('F',np.zeros((5,3,3)),'set to zero').get('F')
|
||||
|
|
Loading…
Reference in New Issue