2022-03-27 02:43:54 +05:30
|
|
|
import sys
|
2021-03-23 19:30:59 +05:30
|
|
|
import random
|
|
|
|
|
2020-06-25 04:07:33 +05:30
|
|
|
import pytest
|
|
|
|
import numpy as np
|
2020-09-28 11:10:43 +05:30
|
|
|
from scipy import stats
|
2021-03-23 19:30:59 +05:30
|
|
|
import h5py
|
2020-09-28 11:10:43 +05:30
|
|
|
|
2020-06-21 02:21:00 +05:30
|
|
|
from damask import util
|
|
|
|
|
|
|
|
class TestUtil:
|
|
|
|
|
2022-03-27 02:43:54 +05:30
|
|
|
@pytest.mark.xfail(sys.platform == 'win32', reason='echo is not a Windows command')
|
2022-03-27 12:33:51 +05:30
|
|
|
def test_run_direct(self):
|
|
|
|
out,err = util.run('echo test')
|
2020-06-21 02:21:00 +05:30
|
|
|
assert out=='test\n' and err==''
|
2020-06-25 04:07:33 +05:30
|
|
|
|
2022-03-27 02:43:54 +05:30
|
|
|
@pytest.mark.xfail(sys.platform == 'win32', reason='echo is not a Windows command')
|
2022-03-27 12:33:51 +05:30
|
|
|
def test_run_env(self):
|
|
|
|
out,err = util.run('sh -c "echo $test_for_execute"',env={'test_for_execute':'test'})
|
2020-06-21 02:21:00 +05:30
|
|
|
assert out=='test\n' and err==''
|
2020-06-25 04:07:33 +05:30
|
|
|
|
2022-03-27 02:43:54 +05:30
|
|
|
@pytest.mark.xfail(sys.platform == 'win32', reason='false is not a Windows command')
|
2022-03-27 12:33:51 +05:30
|
|
|
def test_run_runtime_error(self):
|
2020-11-14 23:54:31 +05:30
|
|
|
with pytest.raises(RuntimeError):
|
2022-03-27 12:33:51 +05:30
|
|
|
util.run('false')
|
2020-11-14 23:54:31 +05:30
|
|
|
|
2022-12-14 00:02:19 +05:30
|
|
|
@pytest.mark.parametrize('input,glue,quote,output',
|
|
|
|
[
|
|
|
|
(None,'',False,'None'),
|
|
|
|
([None,None],'\n',False,'None\nNone'),
|
|
|
|
([-0.5,0.5],'=',False,'-0.5=0.5'),
|
|
|
|
([1,2,3],'_',False,'1_2_3'),
|
|
|
|
([1,2,3],'/',True,'"1"/"2"/"3"'),
|
|
|
|
])
|
|
|
|
def test_srepr(self,input,glue,quote,output):
|
|
|
|
assert output == util.srepr(input,glue,quote)
|
|
|
|
|
2020-06-25 04:07:33 +05:30
|
|
|
@pytest.mark.parametrize('input,output',
|
|
|
|
[
|
2020-09-19 23:08:32 +05:30
|
|
|
([0,-2],[0,-1]),
|
|
|
|
([-0.5,0.5],[-1,1]),
|
2020-06-25 04:07:33 +05:30
|
|
|
([1./2.,1./3.],[3,2]),
|
|
|
|
([2./3.,1./2.,1./3.],[4,3,2]),
|
|
|
|
])
|
|
|
|
def test_scale2coprime(self,input,output):
|
|
|
|
assert np.allclose(util.scale_to_coprime(np.array(input)),
|
|
|
|
np.array(output).astype(int))
|
|
|
|
|
|
|
|
def test_lackofprecision(self):
|
|
|
|
with pytest.raises(ValueError):
|
2020-09-19 23:08:32 +05:30
|
|
|
util.scale_to_coprime(np.array([1/333.333,1,1]))
|
2020-09-28 11:10:43 +05:30
|
|
|
|
|
|
|
|
|
|
|
@pytest.mark.parametrize('rv',[stats.rayleigh(),stats.weibull_min(1.2),stats.halfnorm(),stats.pareto(2.62)])
|
2022-11-24 16:08:16 +05:30
|
|
|
def test_hybridIA_distribution(self,rv):
|
2020-09-28 11:10:43 +05:30
|
|
|
bins = np.linspace(0,10,100000)
|
|
|
|
centers = (bins[1:]+bins[:-1])/2
|
|
|
|
N_samples = bins.shape[0]-1000
|
|
|
|
dist = rv.pdf(centers)
|
|
|
|
selected = util.hybrid_IA(dist,N_samples)
|
|
|
|
dist_sampled = np.histogram(centers[selected],bins)[0]/N_samples*np.sum(dist)
|
|
|
|
assert np.sqrt(((dist - dist_sampled) ** 2).mean()) < .025 and selected.shape[0]==N_samples
|
2020-11-10 01:50:56 +05:30
|
|
|
|
2022-11-24 16:08:16 +05:30
|
|
|
def test_hybridIA_constant(self):
|
|
|
|
N_bins = np.random.randint(20,400)
|
|
|
|
m = np.random.randint(1,20)
|
|
|
|
N_samples = m * N_bins
|
|
|
|
dist = np.ones(N_bins)*np.random.rand()
|
|
|
|
assert np.all(np.sort(util.hybrid_IA(dist,N_samples))==np.arange(N_samples).astype(int)//m)
|
|
|
|
|
|
|
|
def test_hybridIA_linear(self):
|
|
|
|
N_points = np.random.randint(10,200)
|
|
|
|
m = np.random.randint(1,20)
|
|
|
|
dist = np.arange(N_points)
|
|
|
|
N_samples = m * np.sum(dist)
|
|
|
|
assert np.all(np.bincount(util.hybrid_IA(dist*np.random.rand(),N_samples)) == dist*m)
|
|
|
|
|
|
|
|
|
2021-02-28 05:02:53 +05:30
|
|
|
@pytest.mark.parametrize('point,direction,normalize,keepdims,answer',
|
2020-11-10 01:50:56 +05:30
|
|
|
[
|
2021-02-28 05:02:53 +05:30
|
|
|
([1,0,0],'z',False,True, [1,0,0]),
|
|
|
|
([1,0,0],'z',True, False,[1,0]),
|
|
|
|
([0,1,1],'z',False,True, [0,0.5,0]),
|
|
|
|
([0,1,1],'y',True, False,[0.41421356,0]),
|
|
|
|
([1,1,0],'x',False,False,[0.5,0]),
|
|
|
|
([1,1,1],'y',True, True, [0.3660254, 0,0.3660254]),
|
2020-11-10 01:50:56 +05:30
|
|
|
])
|
2021-12-28 15:49:17 +05:30
|
|
|
def test_project_equal_angle(self,point,direction,normalize,keepdims,answer):
|
|
|
|
assert np.allclose(util.project_equal_angle(np.array(point),direction=direction,
|
|
|
|
normalize=normalize,keepdims=keepdims),answer)
|
|
|
|
|
|
|
|
@pytest.mark.parametrize('point,direction,normalize,keepdims,answer',
|
|
|
|
[
|
|
|
|
([1,0,0],'z',False,True, [1,0,0]),
|
|
|
|
([1,0,0],'z',True, False,[1,0]),
|
|
|
|
([0,1,1],'z',False,True, [0,0.70710678,0]),
|
|
|
|
([0,1,1],'y',True, False,[0.5411961,0]),
|
|
|
|
([1,1,0],'x',False,False,[0.70710678,0]),
|
|
|
|
([1,1,1],'y',True, True, [0.45970084,0,0.45970084]),
|
|
|
|
])
|
|
|
|
def test_project_equal_area(self,point,direction,normalize,keepdims,answer):
|
|
|
|
assert np.allclose(util.project_equal_area(np.array(point),direction=direction,
|
|
|
|
normalize=normalize,keepdims=keepdims),answer)
|
2020-11-10 01:50:56 +05:30
|
|
|
|
|
|
|
@pytest.mark.parametrize('fro,to,mode,answer',
|
|
|
|
[
|
|
|
|
((),(1,),'left',(1,)),
|
|
|
|
((1,),(7,),'right',(1,)),
|
|
|
|
((1,2),(1,1,2,2),'right',(1,1,2,1)),
|
|
|
|
((1,2),(1,1,2,2),'left',(1,1,1,2)),
|
|
|
|
((1,2,3),(1,1,2,3,4),'right',(1,1,2,3,1)),
|
|
|
|
((10,2),(10,3,2,2,),'right',(10,1,2,1)),
|
|
|
|
((10,2),(10,3,2,2,),'left',(10,1,1,2)),
|
|
|
|
((2,2,3),(2,2,2,3,4),'left',(1,2,2,3,1)),
|
|
|
|
((2,2,3),(2,2,2,3,4),'right',(2,2,1,3,1)),
|
|
|
|
])
|
|
|
|
def test_shapeshifter(self,fro,to,mode,answer):
|
|
|
|
assert util.shapeshifter(fro,to,mode) == answer
|
|
|
|
|
|
|
|
@pytest.mark.parametrize('fro,to,mode',
|
|
|
|
[
|
|
|
|
((10,3,4),(10,3,2,2),'left'),
|
|
|
|
((2,3),(10,3,2,2),'right'),
|
|
|
|
])
|
|
|
|
def test_invalid_shapeshifter(self,fro,to,mode):
|
|
|
|
with pytest.raises(ValueError):
|
|
|
|
util.shapeshifter(fro,to,mode)
|
|
|
|
|
|
|
|
@pytest.mark.parametrize('a,b,answer',
|
|
|
|
[
|
|
|
|
((),(1,),(1,)),
|
|
|
|
((1,),(),(1,)),
|
|
|
|
((1,),(7,),(1,7)),
|
|
|
|
((2,),(2,2),(2,2)),
|
|
|
|
((1,2),(2,2),(1,2,2)),
|
|
|
|
((1,2,3),(2,3,4),(1,2,3,4)),
|
|
|
|
((1,2,3),(1,2,3),(1,2,3)),
|
|
|
|
])
|
|
|
|
def test_shapeblender(self,a,b,answer):
|
|
|
|
assert util.shapeblender(a,b) == answer
|
2020-11-14 23:54:31 +05:30
|
|
|
|
2020-11-15 15:23:23 +05:30
|
|
|
@pytest.mark.parametrize('style',[util.emph,util.deemph,util.warn,util.strikeout])
|
2020-11-14 23:54:31 +05:30
|
|
|
def test_decorate(self,style):
|
|
|
|
assert 'DAMASK' in style('DAMASK')
|
2021-03-23 19:30:59 +05:30
|
|
|
|
|
|
|
@pytest.mark.parametrize('complete',[True,False])
|
|
|
|
def test_D3D_base_group(self,tmp_path,complete):
|
|
|
|
base_group = ''.join(random.choices('DAMASK', k=10))
|
|
|
|
with h5py.File(tmp_path/'base_group.dream3d','w') as f:
|
2022-03-27 02:39:45 +05:30
|
|
|
f.create_group('/'.join((base_group,'_SIMPL_GEOMETRY')))
|
2021-03-23 19:30:59 +05:30
|
|
|
if complete:
|
2022-03-27 02:39:45 +05:30
|
|
|
f['/'.join((base_group,'_SIMPL_GEOMETRY'))].create_dataset('SPACING',data=np.ones(3))
|
2021-03-23 19:30:59 +05:30
|
|
|
|
|
|
|
if complete:
|
|
|
|
assert base_group == util.DREAM3D_base_group(tmp_path/'base_group.dream3d')
|
|
|
|
else:
|
|
|
|
with pytest.raises(ValueError):
|
|
|
|
util.DREAM3D_base_group(tmp_path/'base_group.dream3d')
|
|
|
|
|
|
|
|
@pytest.mark.parametrize('complete',[True,False])
|
|
|
|
def test_D3D_cell_data_group(self,tmp_path,complete):
|
|
|
|
base_group = ''.join(random.choices('DAMASK', k=10))
|
|
|
|
cell_data_group = ''.join(random.choices('KULeuven', k=10))
|
|
|
|
cells = np.random.randint(1,50,3)
|
|
|
|
with h5py.File(tmp_path/'cell_data_group.dream3d','w') as f:
|
2022-03-27 02:39:45 +05:30
|
|
|
f.create_group('/'.join((base_group,'_SIMPL_GEOMETRY')))
|
|
|
|
f['/'.join((base_group,'_SIMPL_GEOMETRY'))].create_dataset('SPACING',data=np.ones(3))
|
|
|
|
f['/'.join((base_group,'_SIMPL_GEOMETRY'))].create_dataset('DIMENSIONS',data=cells[::-1])
|
2021-03-23 19:30:59 +05:30
|
|
|
f[base_group].create_group(cell_data_group)
|
|
|
|
if complete:
|
2022-03-27 02:39:45 +05:30
|
|
|
f['/'.join((base_group,cell_data_group))].create_dataset('data',shape=np.append(cells,1))
|
2021-03-23 19:30:59 +05:30
|
|
|
|
|
|
|
if complete:
|
|
|
|
assert cell_data_group == util.DREAM3D_cell_data_group(tmp_path/'cell_data_group.dream3d')
|
|
|
|
else:
|
|
|
|
with pytest.raises(ValueError):
|
|
|
|
util.DREAM3D_cell_data_group(tmp_path/'cell_data_group.dream3d')
|
2021-03-31 14:29:21 +05:30
|
|
|
|
|
|
|
|
|
|
|
@pytest.mark.parametrize('full,reduced',[({}, {}),
|
|
|
|
({'A':{}}, {}),
|
|
|
|
({'A':{'B':{}}}, {}),
|
|
|
|
({'A':{'B':'C'}},)*2,
|
|
|
|
({'A':{'B':{},'C':'D'}}, {'A':{'C':'D'}})])
|
2021-04-02 03:03:45 +05:30
|
|
|
def test_prune(self,full,reduced):
|
|
|
|
assert util.dict_prune(full) == reduced
|
2021-03-31 14:29:21 +05:30
|
|
|
|
|
|
|
|
|
|
|
@pytest.mark.parametrize('full,reduced',[({}, {}),
|
|
|
|
({'A':{}}, {}),
|
|
|
|
({'A':'F'}, 'F'),
|
|
|
|
({'A':{'B':{}}}, {}),
|
|
|
|
({'A':{'B':'C'}}, 'C'),
|
|
|
|
({'A':1,'B':2},)*2,
|
|
|
|
({'A':{'B':'C','D':'E'}}, {'B':'C','D':'E'}),
|
|
|
|
({'B':'C','D':'E'},)*2,
|
|
|
|
({'A':{'B':{},'C':'D'}}, {'B':{},'C':'D'})])
|
2021-04-02 03:03:45 +05:30
|
|
|
def test_flatten(self,full,reduced):
|
|
|
|
assert util.dict_flatten(full) == reduced
|
2021-06-02 00:59:35 +05:30
|
|
|
|
|
|
|
|
|
|
|
def test_double_Bravais_to_Miller(self):
|
|
|
|
with pytest.raises(KeyError):
|
|
|
|
util.Bravais_to_Miller(uvtw=np.ones(4),hkil=np.ones(4))
|
|
|
|
|
|
|
|
def test_double_Miller_to_Bravais(self):
|
|
|
|
with pytest.raises(KeyError):
|
|
|
|
util.Miller_to_Bravais(uvw=np.ones(4),hkl=np.ones(4))
|
|
|
|
|
|
|
|
|
|
|
|
@pytest.mark.parametrize('vector',np.array([
|
|
|
|
[1,0,0],
|
|
|
|
[1,1,0],
|
|
|
|
[1,1,1],
|
|
|
|
[1,0,-2],
|
|
|
|
]))
|
|
|
|
@pytest.mark.parametrize('kw_Miller,kw_Bravais',[('uvw','uvtw'),('hkl','hkil')])
|
|
|
|
def test_Miller_Bravais_Miller(self,vector,kw_Miller,kw_Bravais):
|
|
|
|
assert np.all(vector == util.Bravais_to_Miller(**{kw_Bravais:util.Miller_to_Bravais(**{kw_Miller:vector})}))
|
|
|
|
|
|
|
|
@pytest.mark.parametrize('vector',np.array([
|
|
|
|
[1,0,-1,2],
|
|
|
|
[1,-1,0,3],
|
|
|
|
[1,1,-2,-3],
|
|
|
|
[0,0,0,1],
|
|
|
|
]))
|
|
|
|
@pytest.mark.parametrize('kw_Miller,kw_Bravais',[('uvw','uvtw'),('hkl','hkil')])
|
|
|
|
def test_Bravais_Miller_Bravais(self,vector,kw_Miller,kw_Bravais):
|
|
|
|
assert np.all(vector == util.Miller_to_Bravais(**{kw_Miller:util.Bravais_to_Miller(**{kw_Bravais:vector})}))
|
2022-11-19 13:40:00 +05:30
|
|
|
|
|
|
|
|
|
|
|
@pytest.mark.parametrize('extra_parameters',["""
|
|
|
|
p2 : str, optional
|
|
|
|
p2 description 1
|
|
|
|
p2 description 2
|
|
|
|
""",
|
|
|
|
"""
|
|
|
|
|
|
|
|
p2 : str, optional
|
|
|
|
p2 description 1
|
|
|
|
p2 description 2
|
|
|
|
|
|
|
|
""",
|
|
|
|
"""
|
|
|
|
p2 : str, optional
|
|
|
|
p2 description 1
|
|
|
|
p2 description 2
|
|
|
|
"""])
|
|
|
|
@pytest.mark.parametrize('invalid_docstring',["""
|
|
|
|
Function description
|
|
|
|
|
|
|
|
Parameters ----------
|
|
|
|
p0 : numpy.ndarray, shape (...,4)
|
|
|
|
p0 description 1
|
|
|
|
p0 description 2
|
|
|
|
p1 : int, optional
|
|
|
|
p1 description
|
|
|
|
|
|
|
|
Remaining description
|
|
|
|
""",
|
|
|
|
"""
|
|
|
|
Function description
|
|
|
|
|
|
|
|
Parameters
|
|
|
|
----------
|
|
|
|
p0 : numpy.ndarray, shape (...,4)
|
|
|
|
p0 description 1
|
|
|
|
p0 description 2
|
|
|
|
p1 : int, optional
|
|
|
|
p1 description
|
|
|
|
|
|
|
|
Remaining description
|
|
|
|
""",])
|
|
|
|
def test_extend_docstring_parameters(self,extra_parameters,invalid_docstring):
|
|
|
|
test_docstring = """
|
|
|
|
Function description
|
|
|
|
|
|
|
|
Parameters
|
|
|
|
----------
|
|
|
|
p0 : numpy.ndarray, shape (...,4)
|
|
|
|
p0 description 1
|
|
|
|
p0 description 2
|
|
|
|
p1 : int, optional
|
|
|
|
p1 description
|
|
|
|
|
|
|
|
Remaining description
|
|
|
|
"""
|
|
|
|
expected = """
|
|
|
|
Function description
|
|
|
|
|
|
|
|
Parameters
|
|
|
|
----------
|
|
|
|
p0 : numpy.ndarray, shape (...,4)
|
|
|
|
p0 description 1
|
|
|
|
p0 description 2
|
|
|
|
p1 : int, optional
|
|
|
|
p1 description
|
|
|
|
p2 : str, optional
|
|
|
|
p2 description 1
|
|
|
|
p2 description 2
|
|
|
|
|
|
|
|
Remaining description
|
|
|
|
""".split("\n")
|
|
|
|
assert expected == util._docstringer(test_docstring,extra_parameters).split('\n')
|
|
|
|
with pytest.raises(RuntimeError):
|
|
|
|
util._docstringer(invalid_docstring,extra_parameters)
|
|
|
|
|
|
|
|
def test_replace_docstring_return_type(self):
|
|
|
|
class TestClassOriginal:
|
|
|
|
pass
|
|
|
|
|
|
|
|
def original_func() -> TestClassOriginal:
|
|
|
|
pass
|
|
|
|
|
|
|
|
class TestClassDecorated:
|
|
|
|
def decorated_func_bound(self) -> 'TestClassDecorated':
|
|
|
|
pass
|
|
|
|
|
|
|
|
def decorated_func() -> TestClassDecorated:
|
|
|
|
pass
|
|
|
|
|
|
|
|
original_func.__doc__ = """
|
|
|
|
Function description/Parameters
|
|
|
|
|
|
|
|
Returns
|
|
|
|
-------
|
|
|
|
Return value : test_util.TestClassOriginal
|
|
|
|
|
|
|
|
Remaining description
|
|
|
|
"""
|
|
|
|
|
|
|
|
expected = """
|
|
|
|
Function description/Parameters
|
|
|
|
|
|
|
|
Returns
|
|
|
|
-------
|
|
|
|
Return value : test_util.TestClassDecorated
|
|
|
|
|
|
|
|
Remaining description
|
|
|
|
"""
|
|
|
|
assert expected == util._docstringer(original_func,return_type=decorated_func)
|
|
|
|
assert expected == util._docstringer(original_func,return_type=TestClassDecorated.decorated_func_bound)
|