Source code for epsproc.IO

# -*- coding: utf-8 -*-
"""
ePSproc IO functions.
=====================

Module for file IO and data parsing.

Main function: :py:func:`epsproc.IO.readMatEle`:

    readMatEle(fileIn = None, fileBase = None, fType = '.out'):

    Read ePS file(s) and return results as Xarray data structures containing matrix elements.
    File endings specified by fType, default .out.


History
-------
06/11/19        Added jobInfo and molInfo data structures, from ePS file via :py:func:`epsproc.IO.headerFileParse()` and :py:func:`epsproc.IO.molInfoParse()`.
                Still needs a bit of work, and may want to implement other (comp chem) libraries here.

14/10/19        Added/debugged read functions for CrossSecion segments.

27/09/19        Added read functions for EDCS segments.

17/09/19        Added read/write to/from netCDF files for Xarrays.
                Use built-in methods, with work-arounds for complex number format issues.

29/08/19        Updating docs to rst.

26/08/19        Added parsing for E, sym parameters from head of ePS file.
                Added error checking by comparing read mat elements to expected list.
                Changed & fixed Xarray indexing - matrix elements now output with dims (LM, Eke, Sym, mu, it, Type)
                Current code rather ugly however.

19/08/19        Add functions for reading wavefunction files (3D data)

07/08/19        Naming convention tweaks, and some changes to comments, following basic tests with Sphinx.

05/08/19    v1  Initial python version.
                Working, but little error checking as yet. Needs some tidying.


To do
-----
* Add IO for other file segments (only DumpIdy supported so far).
* Better logic & flexibility for file scanning.
* Restructure as class for brevity...?
* More sophisticated methods/data structures for job & molecule info handling.



"""

# Imports
import os
import re
import numpy as np
import pandas as pd
from io import StringIO
import xarray as xr
from datetime import datetime as dt # Import datetime.datetime for now() function

try:
    from pyevtk.hl import gridToVTK
except ImportError as e:
    if e.msg != "No module named 'pyevtk'":
        raise
    print('* pyevtk not found, VTK export not available. ')

# Package fns.
from epsproc.util import matEleSelector, dataGroupSel, matEdimList, BLMdimList, stringRepMap

# ***** Ancillary functions

# File parsing function - scan file for keywords & read segments
#   Following above idiomatic solution, with full IO
#   https://stackoverflow.com/questions/3961265/get-line-number-of-certain-phrase-in-file-python
[docs]def fileParse(fileName, startPhrase = None, endPhrase = None, comment = None, verbose = False): """ Parse a file, return segment(s) from startPhrase:endPhase, excluding comments. Parameters ---------- fileName : str File to read (file in working dir, or full path) startPhrase : str, optional Phrase denoting start of section to read. Default = None endPhase : str, optional Phrase denoting end of section to read. Default = None comment : str, optional Phrase denoting comment lines, which are skipped. Default = None Returns ------- list [lineStart, lineStop], ints for line #s found from start and end phrases. list segments, list of lines read from file. All lists can contain multiple entries, if more than one segment matches the search criteria. """ lineStart = [] # Create empty list to hold line #s lineStop = [] # Create empty list to hold line #s segments = [[]] # Possible to create empty multi-dim array here without knowing # of segments? Otherwise might be easier to use np textscan functions readFlag = False n = 0 # Open file & scan each line. with open(fileName,'r') as f: for (i, line) in enumerate(f): # Note enumerate() here gives lines with numbers, e.g. fullFile=enumerate(f) will read in file with numbers i = i + 1 # Offset for file line numbers (1 indexed) # If line matches startPhrase, print line & append to list. # Note use of lstrip to skip any leading whitespace. # if startPhrase in line: if line.lstrip().startswith(startPhrase): if verbose: print('Found "', startPhrase, '" at line: ', i) lineStart.append(i) readFlag = True # Read lines into segment[] until endPhrase found if readFlag: # Check for end of segment (start of next Command sequence) if endPhrase and ([line.startswith(endP) for endP in endPhrase].count(True) > 0): # This allows for multiple endPhases # NOTE: this will iterate over all chars in a phrase if a single str is passed. # Log stop line and list lineStop.append(i) readFlag = False # Log segment and create next segments.append([]) n += 1 continue # Continue will skip rest of loop # Check for comments, skip line but keep reading elif comment and line.startswith(comment): continue # Continue will skip rest of loop segments[n].append([n, i, line]) # Store line if part of defined segment if verbose: print('Found {0} segments.'.format(n)) return ([lineStart, lineStop], segments) # [:-1])
# Simple wrapper for general fileParse function, ePS dumpIdy segments
[docs]def dumpIdyFileParse(fileName): """ Parse an ePS file for dumpIdy segments. Parameters ---------- fileName : str File to read (file in working dir, or full path) Returns ------- list [lineStart, lineStop], ints for line #s found from start and end phrases. list dumpSegs, list of lines read from file. Lists contain entries for each dumpIdy segment found in the file. """ startPhrase = "DumpIdy - dump" endPhrase = ["+ Command", "Time Now"] # In this case may have multiple end phrases (lines, dumpSegs) = fileParse(fileName, startPhrase, endPhrase) # , '>') # NOTE - with current code dumpSegs has one final blank segment print('Found {0} dumpIdy segments (sets of matrix elements).'.format(len(dumpSegs) - 1)) return lines, dumpSegs[:-1]
# Simple wrapper for general fileParse function, ePS EDCS segments
[docs]def EDCSFileParse(fileName): """ Parse an ePS file for EDCS segments. Parameters ---------- fileName : str File to read (file in working dir, or full path) Returns ------- list [lineStart, lineStop], ints for line #s found from start and end phrases. list dumpSegs, list of lines read from file. Lists contain entries for each dumpIdy segment found in the file. """ startPhrase = "EDCS - differential cross section program" endPhrase = ["+ Command", "Time Now"] # In this case may have multiple end phrases (lines, dumpSegs) = fileParse(fileName, startPhrase, endPhrase) # , '>') # NOTE - with current code dumpSegs has one final blank segment print('Found {0} EDCS segments (sets of scattering results).'.format(len(dumpSegs) - 1)) return lines, dumpSegs[:-1]
# Simple wrapper for general fileParse function, ePS GetCro/CrossSection segments
[docs]def getCroFileParse(fileName): """ Parse an ePS file for GetCro/CrossSection segments. Parameters ---------- fileName : str File to read (file in working dir, or full path) Returns ------- list [lineStart, lineStop], ints for line #s found from start and end phrases. list dumpSegs, list of lines read from file. Lists contain entries for each dumpIdy segment found in the file. """ # startPhrase = "CrossSection - compute photoionization cross section" # For full file segment startPhrase = "COMPOSITE CROSS SECTIONS AT ALL ENERGIES" # Final table of XSect values endPhrase = ["+ Command", "Time Now"] # In this case may have multiple end phrases (lines, dumpSegs) = fileParse(fileName, startPhrase, endPhrase) # , '>') # NOTE - with current code dumpSegs has one final blank segment print('Found {0} CrossSection segments (sets of results).'.format(len(dumpSegs) - 1)) return lines, dumpSegs[:-1]
# Simple wrapper for general fileParse function, check ScatEng and return Eke list
[docs]def scatEngFileParse(fileName): """ Parse an ePS file for ScatEng list. Parameters ---------- fileName : str File to read (file in working dir, or full path) Returns ------- list ekeList, np array of energies set in the ePS file. Lists contain entries for each dumpIdy segment found in the file. """ startPhrase = "ScatEng" endPhrase = "#" comment = "+" (lines, dumpSegs) = fileParse(fileName, startPhrase, endPhrase, comment) # , '>') # Grab E list, assuming just first segment scanned is relevant ekeList = np.genfromtxt(StringIO(dumpSegs[0][0][2][7:])) # print('Expecting {0} energy points.'.format(len(ekeList))) # For np.array, this will fail for singleton dim. print('Expecting {0} energy points.'.format(ekeList.size)) return ekeList
# Simple wrapper for general fileParse function, check symmetries and return list
[docs]def symFileParse(fileName): """ Parse an ePS file for ScatEng list. Parameters ---------- fileName : str File to read (file in working dir, or full path) Returns ------- list symSegs, raw lines from the ePS file. Lists contain entries for each dumpIdy segment found in the file. """ startPhrase = "ScatSym" endPhrase = ["FileName", "\n"] comment = "+" (lines, symSegs) = fileParse(fileName, startPhrase, endPhrase, comment) # , '>') # Grab E list, assuming just first segment scanned is relevant print('Expecting {0} symmetries.'.format(len(symSegs) - 1)) return symSegs[:-1]
# Parse digits from a line using re # https://stackoverflow.com/questions/4289331/how-to-extract-numbers-from-a-string-in-python
[docs]def parseLineDigits(testLine): """ Use regular expressions to extract digits from a string. https://stackoverflow.com/questions/4289331/how-to-extract-numbers-from-a-string-in-python """ return re.findall("[-+]?[.]?[\d]+(?:,\d\d\d)*[\.]?\d*(?:[eE][-+]?\d+)?", testLine)
# ************** Header info parsing functions # Simple wrapper for general fileParse function, extract ePS file header & input, and parse
[docs]def headerFileParse(fileName, verbose = True): """ Parse an ePS file for header & input job info. Parameters ---------- fileName : str File to read (file in working dir, or full path) verbose : bool, default True Print job info from file header if true. Returns ------- jobInfo : dict Dictionary generated from job details. TO DO ----- - Tidy up methods - maybe with parseDigits? - Tidy up dict output. """ startPhrase = "ePolyScat Version" # Read from top of file endPhrase = ["+ End of input reached"] # In this case only have a single end phrases, but need to pass as list to avoid iterating over phrase (lines, dumpSegs) = fileParse(fileName, startPhrase, endPhrase) # , '>') # NOTE - with current code dumpSegs has one final blank segment # print('Read {0} dumpIdy segments (sets of matrix elements).'.format(len(dumpSegs) - 1)) # Parse info to dict - bit ugly, assumes fixed format for start lines # This might be workable if keys are set from ePS source/web list? # jobKeys = ['ePolyScat', 'Authors', 'http', 'cite', 'Starting', 'Using'] # jobInfo = {} # # for line in dumpSegs[0]: # for key in jobKeys: # if line[2].startswith(key): # jobInfo[key] = line[2].strip() # Generate from file directly - good for key:value pairs, but might mangle prose jobInfo = {} jobInfo['comments'] = [] # Keep comments in this list. # Loop over lines, split and sort if possible - UGLY! for n, line in enumerate(dumpSegs[0]): # elements = dumpSegs[0][n][2].strip().split() elements = line[2].strip().split() # print(elements) if len(elements)>0: if elements[0].startswith('#'): jobInfo['comments'].append(line[2].strip()) elif elements[0].startswith('Orb'): jobInfo[elements[0]] = np.asarray(parseLineDigits(dumpSegs[0][n+1][2].strip())).astype('int') # Set next line for OrbOccs else: if len(elements) == 2: jobInfo[elements[0]] = elements[1] # Case for data record value assignments # Check for lines with comments in elif '#' in line[2]: test = line[2].strip().split('#') # Split at comment testStart = test[0].split() if len(testStart) == 2: jobInfo[testStart[0]] = testStart[1] # Case for data record assignments # This might be redundant... but keep for safety. else: jobInfo[elements[0]] = line[2].strip().split('#') # For other cases keep full line, split at comments # For all other cases keep full line, split at comments else: jobInfo[elements[0]] = line[2].strip().split('#') # Print jobInfo if verbose: print('*** Job info from file header.\n') [print(line.strip('#')) for line in jobInfo['comments'][0:4]] return jobInfo
# Simple wrapper for general fileParse function, extract ePS molecular info, and parse
[docs]def molInfoParse(fileName, verbose = True): """ Parse an ePS file for header & input job info. Parameters ---------- fileName : str File to read (file in working dir, or full path) verbose : bool, default True Print job info from file header if true. Returns ------- molInfo : dict Dictionary with atom & orbital details. """ # Extract mol segment startPhrase = "Selecting orbitals" # Read from top of file endPhrase = ["+ "] # In this case only have a single end phrases, but need to pass as list to avoid iterating over phrase (lines, dumpSegs) = fileParse(fileName, startPhrase, endPhrase) # Basic parsing to lists orbList = [] atomList = [] for line in dumpSegs[0]: if line[2].startswith('Selecting'): orbList.append(line[2].strip().lstrip('Selecting ')) if line[2].startswith('Z'): atomList.append(line[2].strip()) orbList = orbList[1:] if verbose: print('\n*** Found orbitals') print(*orbList, sep='\n') print('\n*** Found atoms') print(*atomList, sep='\n') # Sort orbs to np.array orbTable = [] [orbTable.append(parseLineDigits(orb)) for orb in orbList] orbTable = np.asarray(orbTable).astype('float') # Sort coords to np.array atomTable = [] [atomTable.append(parseLineDigits(atom)) for atom in atomList] atomTable = np.asarray(atomTable).astype('float') # Compile to dict molInfo = {'atomList':atomList, 'atomTable':atomTable, 'orbList':orbList, 'orbTable':orbTable} return molInfo
# ************* DumpIdy parsing # Parse a DumpIdy segment #TODO: More attribs, and convert attribs to dict, or other more structured format. #TODO: Error checking for missing matrix elements or IO issues - see Matlab code. #TODO: Convert attribs to dict for simpler assignments later
[docs]def dumpIdySegParse(dumpSeg): """ Extract values from dumpIdy file segments. Parameters ---------- dumpSeg : list One dumpIdy segment, from dumpSegs[], as returned by :py:func:`epsproc.IO.dumpIdyFileParse()` Returns ------- np.array rawIdy, array of matrix elements, [m,l,mu,ip,it,Re,Im] list attribs, list [Label, value, units] Notes ----- Currently this is a bit messy, and relies on fixed DumpIDY format. No error checking as yet. Not yet reading all attribs. Example ------- >>> matEle, attribs = dumpIdySegParse(dumpSegs[0]) """ # Use lists to collect data, and convert format at the end attribs = [] rawIdy = [] # print(len(dumpSeg)) # Parse data block # Use native python, or np.genfromtxt for testLine in dumpSeg[12:-1]: # vals = parseLineDigits(testLine[2]) # lineOut = [] # [lineOut.append(float(val)) for val in vals] # Convert to float # rawIdy.append(lineOut) tmp=np.genfromtxt(StringIO(testLine[2])) rawIdy.append(tmp) # Parse header lines (structured, selected by line #) attribs.append(['E', np.float(parseLineDigits(dumpSeg[3][2])[0]), 'eV']) attribs.append(['Ehv', np.float(parseLineDigits(dumpSeg[4][2])[0]), 'eV']) SF = np.genfromtxt(parseLineDigits(dumpSeg[5][2])) SF = SF[0] + SF[1]*1j attribs.append(['SF', SF, 'sqrt(MB)']) attribs.append(['Lmax', np.int(parseLineDigits(dumpSeg[10][2])[0]), '']) # Parse symmetries - multiple .spilt(), or use re? # attribs.append(['Syms', parseLineDigits(dumpSeg[7][2]), '']) symList = dumpSeg[7][2].split() # Append as nested lists # sym = [] # [sym.append([symList[n], symList[n+1].strip('=')]) for n in range(1,6,2)] # attribs.append(['Syms', sym, '']) # Append as indendent lists [attribs.append([symList[n], symList[n+1].strip('=')]) for n in range(1,6,2)] attribs.append(['QNs', dumpSeg[11][2].split(), '']) return np.asarray(rawIdy), attribs
# Functional form for parsing full set of mat elements and putting in xarray
[docs]def dumpIdySegsParseX(dumpSegs, ekeListUn, symSegs): """ Extract data from ePS dumpIdy segments into usable form. Parameters ---------- dumpSegs : list Set of dumpIdy segments, i.e. dumpSegs, as returned by :py:func:`epsproc.IO.dumpIdyFileParse()` ekeListUn : list List of energies, used for error-checking and Xarray rearraging, as returned by :py:func:`epsproc.IO.scatEngFileParse()` Returns ------- xr.array Xarray data array, containing matrix elements etc. Dimensions (LM, Eke, Sym, mu, it, Type) int Number of blank segments found. Example ------- >>> data = dumpIdySegsParseX(dumpSegs) """ dataList = [] ekeList = [] blankSegs = 0 # Loop over DumpIdy segments, extract data & reformat # If blank, skip parser and append blankSegs. for dumpSeg in dumpSegs: if len(dumpSeg)>6: segBlock, attribs = dumpIdySegParse(dumpSeg) dataList.append([segBlock[:,0:5].T, segBlock[:,5]+1j*segBlock[:,6], attribs]) # Switch l,m - with advanced indexing, other methods faster...? https://stackoverflow.com/questions/4857927/swapping-columns-in-a-numpy-array # Now done later via pd.MultiIndex.swaplevel() # dataList[-1][0][[0,1],:] = dataList[-1][0][[1,0],:] # dataList.append([segBlock[:,0:5].T, segBlock[:,5]+1j*segBlock[:,6], attribs]) # dataList.append([segBlock[:,0:5], segBlock[:,5]+1j*segBlock[:,6], attribs]) ekeList.append(attribs[0][1]) else: blankSegs += 1 ekeList.append(np.nan) # Check energies vs. input list, and number of symmetries ekeTest = np.unique(ekeList) if ekeTest.size != ekeListUn.size: print("*** Warning: Found {0} energies, expected {1}".format(ekeTest.size,ekeListUn.size)) # Check here according to expected input, but should also be logged in blankSegs above. if len(ekeList) != ekeTest.size * len(symSegs): print("*** Warning: Missing records, expected {0}, found {1}.".format(ekeTest.size * len(symSegs),len(ekeList))) #**** Convert to xarray - ugly loop version, probably a better way to do this! #TODO Should: # - be able to loop more cleanly over attribs - set as dict? # - integrate with above loop # - compile into dataSet or dataArray more directly...? # - Check and compile against Eke list (muliple symetries per Eke), and keep this as a separate coord. dataArrays = [] dataSym = [] ekeVal = ekeList[0] print('\nProcessing segments to Xarrays...') eLoop = 0 # For logging E vs. sym segments and looping output. for n, data in enumerate(dataList): attribs = data[2] # V1 - EASIEST WAY, but leads to dimensional issues later! #TODO: consider setting mu as a separate dim. Maybe also (ip,it)...? # QNs = pd.MultiIndex.from_arrays(data[0].astype('int8'), names = attribs[-1][1][0:-1]) # QNs = QNs.swaplevel(0, 1) # Switch l,m indexes # Esyms = pd.MultiIndex.from_arrays([np.array(attribs[0][1]), [attribs[3][1], attribs[5][1]]], names=['E', 'Sym']) # pd.MultiIndex.from_tuples([(np.array(attribs[0][1]), [attribs[3][1], attribs[5][1]])], names=['E', 'Sym']) # Esyms = pd.MultiIndex.from_tuples([(attribs[0][1],attribs[4][1],attribs[5][1],attribs[6][1])],names=[attribs[0][0],attribs[4][0],attribs[5][0],attribs[6][0]]) #dataArrays.append(xr.DataArray(data[1], coords={'ES': Esyms, 'LM':QNs}, dims = ['ES','LM'])) # AH - issue is number of labels - can't lable singleton dim it seems, but can expand #TODO: consider setting E as a separate dim, will be singleton for each set of syms. Might make more sense for later manipulations (sum over sym or E). # tmp = xr.DataArray(np.asarray(data[1]), coords={'LM':QNs}, dims = ['LM']) # tmp = tmp.expand_dims({'Sym':Syms, 'Eke':[attribs[0][1]]}) # This is OK, but still ties Eke and Sym coords (same number of elements) # # tmp = tmp.expand_dims({'Sym':Syms}) # # v2 - separate dims, (LM, E, Syms, ip, it) # LM = pd.MultiIndex.from_arrays(data[0][0:2,:].astype('int8'), names = attribs[-1][1][0:2]) # LM = LM.swaplevel(0, 1) # Switch l,m indexes # mu = data[0][2,:] # ip = data[0][3,:] # it = data[0][4,:] # Syms = pd.MultiIndex.from_tuples([(attribs[4][1],attribs[5][1],attribs[6][1])],names=[attribs[4][0],attribs[5][0],attribs[6][0]]) # # tmp = xr.DataArray(np.asarray(data[1]), coords={'LM':LM, 'mu':mu, 'ip':ip, 'it':it}, dims = ['LM','mu','it','ip']) # This works... but still keeps full lenght for additional label coords. # Should be able to reduce on these... but can't work out how (tried groupby etc.) # tmp = xr.DataArray(np.asarray(data[1]), coords={'LM':LM}, dims = ['LM']) # tmp = tmp.expand_dims({'Sym':Syms, 'Eke':[attribs[0][1]]}) # tmp = tmp.expand_dims({'mu':mu, 'ip':ip, 'it':it}) # *** v2b - assign as v1/v2, then sort Xarrays before restacking # This works, but note assumption of stacking order (E, then Syms) QNs = pd.MultiIndex.from_arrays(data[0].astype('int8'), names = attribs[-1][1][0:-1]) QNs = QNs.swaplevel(0, 1) # Switch l,m indexes Syms = pd.MultiIndex.from_tuples([(attribs[4][1],attribs[5][1],attribs[6][1])],names=[attribs[4][0],attribs[5][0],attribs[6][0]]) # Original code - set according to LM, then expand dims. tmp = xr.DataArray(np.asarray(data[1]), coords={'LM':QNs}, dims = ['LM']) tmp = tmp.expand_dims({'Sym':Syms, 'Eke':[attribs[0][1]]}) # New code 24/10/19 - set all coords, including non-dim coords such as SF and Ehv # Should be able to set in one call... but get dim issues here, probably with multiple singletons. # tmp = xr.DataArray(np.asarray(data[1]), coords={'LM':QNs, 'Sym':Syms, 'Eke':[attribs[0][1]], 'Ehv':[attribs[1][1]], 'SF':[attribs[2][1]]}, dims = ['LM', 'Sym', 'Eke']) # Setting as per previous code, then adding addtional singleton non-dimensional coords seems to be OK however... tmp['Ehv']=(('Eke'),[attribs[1][1]]) # Link to single dim... should be OK? # tmp['SF']=(('Eke'),[attribs[2][1]]) # Link to single dim... should be OK? tmp['SF']=(('Eke','Sym'),np.array(attribs[2][1]).reshape(1,1)) # Link to Eke & sym (i.e. single set of matE) # Assign any other attributes - note that some attributes may be dropped when combining arrays below for a in attribs: tmp.attrs[a[0]] = a[1] # Currently set without units, multiple values here give combine issues below. # dataArrays.append(tmp) # tmp = matEleGroupDimX(tmp) # Broken? dataArrays.append(matEleGroupDimX(tmp)) #TODO: Fix scale-factor propagation... this is currently dropped as an inconsistent attrib value. # Something like: dataSF.append(dataArrays[-1].SF) # dataArrays.append(matEleGroupDimXnested(tmp.copy())) # Broken...? # dataArrays.append(tmp) #TODO: UGLY - need to check and combine according to number of Syms (check matlab code) # ALSO NOT WORKING PROPERLY - might be issue with equality? # BETTER TO JUST SET 2D arrays here! # if n == (len(ekeList)/2 - 1): # dataSym.append(xr.combine_nested(dataArrays, concat_dim=['Eke'])) # dataArrays = [] # # if n == len(ekeList)-1: # dataSym.append(xr.combine_nested(dataArrays, concat_dim=['Eke'])) # # if n == (len(ekeList)/2 - 1): # # dataSym.append(xr.combine_nested(dataArrays, concat_dim=['Eke'])) # dataArrays1 = dataArrays.copy() # dataArrays = [] # if n == (len(ekeList)/2 - 1): # dataSym.append(dataArrays) # dataArrays = [] # # if n == len(ekeList)-1: # dataSym.append(dataArrays) # if (n > 1) and (attribs[0][1] == ekeVal): # print('Found {0} energies'.format(n+1)) # dataSym.append(xr.combine_nested(dataArrays, concat_dim=['Eke'])) # Now with loop on known energy list. if eLoop == ekeListUn.size-1: dataSym.append(xr.combine_nested(dataArrays, concat_dim=['Eke'])) dataArrays = [] eLoop = 0 else: eLoop = eLoop+1 # **** v3 manually sort data first... # SEE CODE BELOW, matEleGroupDim() # Assign any other attributes - note that some attributes may be dropped when combining arrays below # for a in attribs: # tmp.attrs[a[0]] = a[1] # Currently set without units, multiple values here give combine issues below. # # dataArrays.append(tmp) # Stack by syms (per eke) as necessary # if ekeList[n] == ekeVal: # # daOut = xr.combine_nested(dataArrays, concat_dim=['Sym']) # dataArrays.append(tmp) # else: # dataSym.append(xr.combine_nested(dataArrays, concat_dim=['Sym'])) # dataArrays = [] # dataArrays.append(tmp) # ekeVal = ekeList[n] # Combine to single xarray # Note xarray > v0.12.1 # daOut = xr.combine_nested(dataSym, concat_dim=['Sym']) daOut = xr.combine_nested(dataSym, concat_dim=['Sym']) # daOut = xr.combine_nested([dataArrays1, dataArrays], concat_dim=['Sym','Eke']) # daOut = xr.combine_nested(np.array(dataArrays), concat_dim=['Sym','Eke']) # Dim issues here - try np.array... NOPE. # daOut = xr.combine_nested(dataArrays, concat_dim=[None]) # daOut = xr.merge(dataArrays) # daOut = xr.combine_by_coords(dataArrays) # daOut = dataArrays # daOut = daOut.expand_dims({'Eke':[attribs[0][1]]}) # # v3 - Sort data before putting into Xarray # # NOW REPLACED ABOVE by sorting of Xarrays - code here may be faster, but less robust. # dataArrays = [] # for data in dataList: # attribs = data[2] # tmp = matEleGroupDim(data) # dataArrays.append(tmp) # # # Can only concat over it at the moment due to duplicate values issue # # THIS WILL BREAK LATER!!!! # # TODO: fix use of Xarrays and dimension issues. # daOut = xr.combine_nested(dataArrays, concat_dim=['it']) # Set any other global attribs daOut.attrs['dataType'] = 'matE' # Set dataType for use later. # SF testing vs. symmetry - remove Sym dim if not necessary, for easier computation later! # Added 25/10/19, not yet well tested. if daOut.SF.diff('Sym').max().pipe(np.abs) < np.finfo(complex).eps: daOut['SF'] = ('Eke', daOut.SF.values[:,0]) return daOut.transpose(), blankSegs # NOTE transpose to reverse ordering of dims.
# ************* EDCS parsing # Parse a EDCS segment (roughly)
[docs]def EDCSSegParse(dumpSeg): """ Extract values from EDCS file segments. Parameters ---------- dumpSeg : list One EDCS segment, from dumpSegs[], as returned by :py:func:`epsproc.IO.EDCSFileParse()` Returns ------- np.array EDCS, array of scattering XS, [theta, Cross Section (Angstrom^2)] list attribs, list [Label, value, units] Notes ----- Currently this is a bit messy, and relies on fixed EDCS format. No error checking as yet. Not yet reading all attribs. Example ------- >>> EDCS, attribs = EDCSSegParse(dumpSegs[0]) """ # Use lists to collect data, and convert format at the end attribs = [] EDCS = [] attribs.append(['E', np.float(parseLineDigits(dumpSeg[13][2])[0]), 'eV']) # For each line convert to float - bit ugly, but works for testLine in dumpSeg[67:]: EDCS.append(np.genfromtxt(StringIO(testLine[2]))) return np.asarray(EDCS), attribs
# Functional form for parsing full set of mat elements and putting in xarray
[docs]def EDCSSegsParseX(dumpSegs): """ Extract data from ePS EDCS segments into usable form. Parameters ---------- dumpSegs : list Set of dumpIdy segments, i.e. dumpSegs, as returned by :py:func:`epsproc.IO.EDCSFileParse()` Returns ------- xr.array Xarray data array, containing cross sections. Dimensions (Eke, theta) int Number of blank segments found. (CURRENTLY not implemented.) Example ------- >>> data = EDCSSegsParseX(dumpSegs) Notes ------ A rather cut-down version of :py:func:`epsproc.IO.dumpIdySegsParseX()`, no error checking currently implemented. """ dataList = [] dataArray = [] ekeList = [] blankSegs = 0 # Loop over DumpIdy segments, extract data & reformat # If blank, skip parser and append blankSegs. for dumpSeg in dumpSegs: if len(dumpSeg)>4: segBlock, attribs = EDCSSegParse(dumpSeg) dataArray.append(segBlock[:,1]) # For brevity, just stack XS data here - will save Xarray sorting later. dataList.append([segBlock[:,0], segBlock[:,1], attribs]) ekeList.append(attribs[0][1]) else: blankSegs += 1 ekeList.append(np.nan) # Dump lists into Xarray - will work provided same theta over all records. daOut = xr.DataArray(np.asarray(dataArray), coords={'E':ekeList, 'Theta':segBlock[:,0]}, dims = ['E','Theta']) daOut.attrs['dataType'] = 'EDCS' # Set dataType for use later. # Set units - should set from file ideally. daOut.attrs['units'] = 'Angs^2' daOut.E.attrs['units'] = 'eV' daOut.Theta.attrs['units'] = 'deg.' return daOut, blankSegs
# ************* CrossSection parsing # Basically same as EDCS, except read a table of values per symmetry. # Parse a getCro/CrossSection segment (roughly)
[docs]def getCroSegParse(dumpSeg): """ Extract values from GetCro/CrossSection file segments. Parameters ---------- dumpSeg : list One CrossSection segment, from dumpSegs[], as returned by :py:func:`epsproc.IO.getCroFileParse()` Returns ------- np.array CrossSections, table of results vs. energy. list attribs, list [Label, value, units] Notes ----- Currently this is a bit messy, and relies on fixed CrossSection output format. No error checking as yet. Not yet reading all attribs. Example ------- >>> XS, attribs = getCroSegParse(dumpSegs[0]) """ # Use lists to collect data, and convert format at the end # attribs = [] XS = [] # attribs.append(['E', np.float(parseLineDigits(dumpSeg[13][2])[0]), 'eV']) # attribs.append(dumpSeg[1][2].split()) # Set header line attribs = dumpSeg[1][2].split() # Set header line # For each line convert to float - bit ugly, but works for testLine in dumpSeg[2:]: XS.append(np.genfromtxt(StringIO(testLine[2]))) return np.asarray(XS), attribs
# Functional form for parsing full set of mat elements and putting in xarray
[docs]def getCroSegsParseX(dumpSegs, symSegs, ekeList): """ Extract data from ePS getCro/CrossSecion segments into usable form. Parameters ---------- dumpSegs : list Set of dumpIdy segments, i.e. dumpSegs, as returned by :py:func:`epsproc.IO.getCroFileParse()` Returns ------- xr.array Xarray data array, containing cross sections. Dimensions (Eke, theta) int Number of blank segments found. (CURRENTLY not implemented.) Example ------- >>> data = getCroSegsParseX(dumpSegs) Notes ------ A rather cut-down version of :py:func:`epsproc.IO.dumpIdySegsParseX()`, no error checking currently implemented. """ dataList = [] dataArray = [] #ekeList = [] blankSegs = 0 # Loop over DumpIdy segments, extract data & reformat # If blank, skip parser and append blankSegs. for n, dumpSeg in enumerate(dumpSegs): if len(dumpSeg)>1: segBlock, attribs = getCroSegParse(dumpSeg) # Create Xarray - basic # daTmp = xr.DataArray(np.asarray(segBlock[:,2:]), # coords={'Ehv':segBlock[:,1], 'XC data':attribs[1:-1:2]}, # 'Type':attribs[2:-1:2]}, # dims = ['Ehv', 'XC data']) #, 'Type']) # Create Xarray - set MultiIndex first & then rearrange. tList = [c[0] for c in attribs[2::2]] # Set type as char typesPD = pd.MultiIndex.from_arrays([attribs[1:-1:2], tList], names = ['XC', 'Type']) daTmp = xr.DataArray(np.asarray(segBlock[:,2:]), coords={'Ehv':segBlock[:,1], 'Ctype':typesPD}, dims = ['Ehv', 'Ctype']).unstack() # Add singleton dim and store. # Note this assumes len(symSegs) >= len(dumpSegs) # Symmetries are not listed in getCro output. sRep = {'ScatSym':'Total', 'ScatContSym':'Cont'} # Dicitonary look-up for sym names try: # dataList.append(daTmp.expand_dims({'Sym':symSegs[n]})) symList = [symString[2].split()[1][1:3] for symString in symSegs[n]] symRep = [stringRepMap(symString[2].split()[0],sRep) for symString in symSegs[n]] except IndexError as e: if e.args[0] != 'list index out of range': raise # dataList.append(daTmp.expand_dims({'Sym':'Missing'})) symList = ['All', 'All'] symRep = ['Total', 'Cont'] # Set MultiIndex for syms Syms = pd.MultiIndex.from_arrays([[symList[0]], [symList[1]]], names=symRep) # Works as expected, rather ugly! dataList.append(daTmp.expand_dims({'Sym':Syms})) else: blankSegs += 1 #ekeList.append(np.nan) # Stack lists by symmetry - this is currently assumed from symSegs # if len(dumpSegs) == len(symSegs): daOut = xr.combine_nested(dataList, concat_dim = ['Sym']) daOut.attrs['dataType'] = 'XSect' # Set dataType for use later. # Set units - should set from file ideally. daOut.Ehv.attrs['units'] = 'eV' daOut.attrs['units'] = 'Mb' # Reset energies to Eke, and shift key dim - might be a simpler/shorter way to do this...? daOut['EhvOrig'] = daOut['Ehv'] daOut['Ehv'] = ekeList daOut = daOut.rename({'Ehv':'Eke', 'EhvOrig':'Ehv'}) # daOut.rename({'Ehv':'Eke'}) # daOut.rename({'EhvOrig':'Ehv'}) return daOut, blankSegs
# ************* MatEle parsing/sorting - needs a tidy up. # Linear version of code, for very specific cases. # Linear tree ip > mu >it # UGLY... also not working now? May have passed Xarray dataset in testing by mistake?
[docs]def matEleGroupDimX(daIn): """ Group ePS matrix elements by redundant labels (Xarray version). Group by ['ip', 'it', 'mu'] terms, all have only a few values. Rename 'ip':1,2 as 'Type':'L','V' TODO: better ways to do this? Via Stack/Unstack? http://xarray.pydata.org/en/stable/api.html#id16 See also tests in funcTests_210819.py for more versions/tests. Parameters ---------- data : Xarray Data array with matrix elements to be split and recombined by dims. Returns ------- data : Xarray Data array with reordered matrix elements (dimensions). """ daRedList = [] daRed = daIn # Split on mu - values from set {-1,0,1} depending on symmetry daRedList = [] uVals = np.unique(daRed.mu) for n, val in enumerate(uVals): tmp = matEleSelector(daRed, inds = {'mu':val}) tmp = tmp.expand_dims({'mu':[val]}) daRedList.append(tmp) # Restack daRed = xr.combine_nested(daRedList, concat_dim = 'mu') # Split on it daRedList = [] uVals = np.unique(daRed.it) for n, val in enumerate(uVals): tmp = matEleSelector(daRed, inds = {'it':val}) tmp = tmp.expand_dims({'it':[val]}) daRedList.append(tmp) # Restack daRed = xr.combine_nested(daRedList, concat_dim = 'it') # Split on 'ip' - will always be (1,2), and split matEle into two ipLabel = ['L','V'] daRedList = [] for n, val in enumerate(range(1,3)): tmp = matEleSelector(daRed, inds = {'ip':val}) tmp = tmp.expand_dims({'Type':[ipLabel[n]]}) daRedList.append(tmp) # Restack daRed = xr.combine_nested(daRedList, concat_dim = 'Type') return daRed
# Subselections using matEleSelector # THIS IS UGLY, but seems to work consistently - get da of correct dims out (with multilevel coords in). # Could also try dataset to array for split and recombine...? # http://xarray.pydata.org/en/v0.12.3/reshaping.html
[docs]def matEleGroupDimXnested(da): """ Group ePS matrix elements by redundant labels (Xarray version). Group by ['ip', 'it', 'mu'] terms, all have only a few values. TODO: better ways to do this? See also tests in funcTests_210819.py for more versions/tests. Parameters ---------- data : Xarray Data array with matrix elements to be split and recombined by dims. """ indList = ['ip','it','mu'] daRedList = [] for x in np.unique(da[indList[0]]): daRedList0 = [] for y in np.unique(da[indList[1]]): daRedList1 = [] for z in np.unique(da[indList[2]]): red = matEleSelector(da, inds = {indList[0]:x, indList[1]:y, indList[2]:z}) red = red.expand_dims({indList[0]:[x], indList[1]:[y], indList[2]:[z]}) daRedList1.append(red) daOut1 = xr.combine_nested(daRedList1, concat_dim = indList[2]) daRedList0.append(daOut1) daOut2 = xr.combine_nested(daRedList0, concat_dim = indList[1]) daRedList.append(daOut2) daOut = xr.combine_nested(daRedList, concat_dim = indList[0]) return daOut
# UGH THIS IS SO UGLY, please make it better. # Should be a neat recursive tree method here, probably also something canned! # Or with native Xarray functionality, but in testing couldn't get this to work properly.
[docs]def matEleGroupDim(data, dimGroups = [3, 4, 2]): """ Group ePS matrix elements by redundant labels. Default is to group by ['ip', 'it', 'mu'] terms, all have only a few values. TODO: better ways to do this? Shoud be possible at Xarray level. Parameters ---------- data : list Sections from dumpIdy segment, as created in dumpIdySegsParseX() Ordering is [labels, matElements, attribs]. """ # Basic version assuming dims # # Split on ip dataIP = [] dInd = 3 dataIP = dataGroupSel(data, dInd) # Split on IT dInd = 4 dataIT = [] for dataSub in dataIP: temp = dataSub dataIT.extend(dataGroupSel(dataSub, dInd)) # Split on mu dInd = 2 dataMU = [] for dataSub in dataIT: dataMU.extend(dataGroupSel(dataSub, dInd)) # Put into Xarray (code from dumpIdySegsParseX) # Label singleton dims directly, then stack. dataArrays = [] attribs = data[2] # Shared attribs for dataSub in dataMU: LM = pd.MultiIndex.from_arrays(dataSub[0][0:2,:].astype('int8'), names = attribs[-1][1][0:2]) LM = LM.swaplevel(0, 1) # Switch l,m indexes mu = [dataSub[0][2,0].astype('int8')] # Already set to single values above. ip = [dataSub[0][3,0].astype('int8')] it = [dataSub[0][4,0].astype('int8')] Syms = pd.MultiIndex.from_tuples([(attribs[4][1],attribs[5][1],attribs[6][1])],names=[attribs[4][0],attribs[5][0],attribs[6][0]]) #dataArrays.append(xr.DataArray(data[1], coords={'ES': Esyms, 'LM':QNs}, dims = ['ES','LM'])) # AH - issue is number of labels - can't lable singleton dim it seems, but can expand #TODO: consider setting E as a separate dim, will be singleton for each set of syms. Might make more sense for later manipulations (sum over sym or E). # tmp = xr.DataArray(np.asarray(data[1]), coords={'LM':QNs}, dims = ['LM']) # tmp = tmp.expand_dims({'Sym':Syms, 'Eke':[attribs[0][1]]}) # This is OK, but still ties Eke and Sym coords (same number of elements) # # tmp = tmp.expand_dims({'Sym':Syms}) # tmp = xr.DataArray(np.asarray(data[1]), coords={'LM':LM, 'mu':mu, 'ip':ip, 'it':it}, dims = ['LM','mu','it','ip']) tmp = xr.DataArray(np.asarray(dataSub[1]), coords={'LM':LM}, dims = ['LM']) tmp = tmp.expand_dims({'Sym':Syms, 'Eke':[attribs[0][1]]}) tmp = tmp.expand_dims({'mu':mu, 'ip':ip, 'it':it}) # Assign any other attributes - note that some attributes may be dropped when combining arrays below for a in attribs: tmp.attrs[a[0]] = a[1] # Currently set without units, multiple values here give combine issues below. dataArrays.append(tmp) # Recombine along it da = xr.combine_nested(dataArrays, concat_dim = ['it']) return da
# Function for grabbing files or scanning dir for files. # Note raw string for docstring as one method of keeping raw string in example.
[docs]def getFiles(fileIn = None, fileBase = None, fType = '.out'): r""" Read ePS file(s) and return results as Xarray data structures. File endings specified by fType, default .out. Parameters ---------- fileIn : str, list of strs, optional. File(s) to read (file in working dir, or full path). Defaults to current working dir if only a file name is supplied. For consistent results, pass raw strings, e.g. ``fileIn = r"C:\share\code\ePSproc\python_dev\no2_demo_ePS.out"`` fileBase : str, optional. Dir to scan for files. Currently only accepts a single dir. Defaults to current working dir if no other parameters are passed. fType : str, optional File ending for ePS output files, default '.out' Returns ------- list List of Xarray data arrays, containing matrix elements etc. from each file scanned. """ currDir = os.getcwd() if fileBase is None: fileBase = currDir if fileIn is not None: # Wrap in list if only single file passed if type(fileIn) is str: fileIn = [fileIn] fList = [] for file in fileIn: # Check file & path are valid fTest = os.path.split(file) if not fTest[0]: fList.append(os.path.join(currDir, file)) else: fList.append(file) # Display message print('\n*** Scanning file(s)') print(fList) else: # Filenames only # fList = [f for f in os.listdir(fileBase) if f.endswith(fType)] # With full path fList = [os.path.join(fileBase, f) for f in os.listdir(fileBase) if f.endswith(fType)] # Display message print('\n*** Scanning dir') print(fileBase) print('Found {0} {1} file(s)\n'.format(len(fList), fType)) return fList
#****** Master function to read a file, or dir, of ePS outputs. # Some of the logic and methods could do with a revisit/tidy-up here... #TODO: Add error checking on paths using os.path.isdir/isfile etc. #TODO: Check/fix paths if incorrectly passed, e.g. https://stackoverflow.com/a/21605790 # ADDED: type switch for matEle or EDCS, probably more to come. Should rename function!
[docs]def readMatEle(fileIn = None, fileBase = None, fType = '.out', recordType = 'DumpIdy'): r""" Read ePS file(s) and return results as Xarray data structures. File endings specified by fType, default *.out. Parameters ---------- fileIn : str, list of strs, optional. File(s) to read (file in working dir, or full path). Defaults to current working dir if only a file name is supplied. For consistent results, pass raw strings, e.g. ``fileIn = r"C:\share\code\ePSproc\python_dev\no2_demo_ePS.out"`` fileBase : str, optional. Dir to scan for files. Currently only accepts a single dir. Defaults to current working dir if no other parameters are passed. fType : str, optional File ending for ePS output files, default '.out' recordType : str, optional, default 'DumpIdy' Type of record to scan for, currently set for 'DumpIdy', 'EDCS' or 'CrossSection'. For a full list of descriptions, types and sources, run: >>> epsproc.util.dataTypesList() Returns ------- list List of Xarray data arrays, containing matrix elements etc. from each file scanned. Examples -------- >>> dataSet = readMatEle() # Scan current dir >>> fileIn = r'C:\share\code\ePSproc\python_dev\no2_demo_ePS.out' >>> dataSet = readMatEle(fileIn) # Scan single file >>> dataSet = readMatEle(fileBase = r'C:\share\code\ePSproc\python_dev') # Scan dir .. note:: * Files are scanned for matrix element output flagged by "DumpIdy" headers. * Each segment found is parsed for attributes and data (set of matrix elements). * Matrix elements and attributes are combined and output as an Xarray array. """ # Set files to read, either: # - dir (one only) to scan using os # - file(s) and add to list with full path # - default case is scan current working dir. print('*** ePSproc readMatEle(): scanning files for ' + recordType + ' segments.') # Call function to check files or scan dir. fList = getFiles(fileIn = fileIn, fileBase = fileBase, fType = fType) # Loop over fList and scan ePS files dataSet = [] for file in fList: print('\n*** Reading ePS output file: ', file) # Scan the file and parse segments #lines, dumpSegs = dumpIdyFileParse(os.path.join(fileBase, file)) if recordType is 'DumpIdy': ekeList = scatEngFileParse(file) symSegs = symFileParse(file) print('Scanning CrossSection segments.') print('Expecting {0} DumpIdy segments.'.format(ekeList.size * len(symSegs))) lines, dumpSegs = dumpIdyFileParse(file) data, blankSegs = dumpIdySegsParseX(dumpSegs, ekeList, symSegs) if recordType is 'EDCS': # print('Expecting {0} EDCS segments.'.format(ekeList.size)) print('Scanning EDCS segments.') lines, dumpSegs = EDCSFileParse(file) data, blankSegs = EDCSSegsParseX(dumpSegs) # , ekeList, symSegs) if recordType is 'CrossSection': ekeList = scatEngFileParse(file) symSegs = symFileParse(file) print('Scanning CrossSection segments.') print('Expecting {0} CrossSection segments.'.format(len(symSegs)+1)) # Assuming 1 segment per symmetry, plus symm-summed case. lines, dumpSegs = getCroFileParse(file) data, blankSegs = getCroSegsParseX(dumpSegs, symSegs, ekeList) # Add some additional properties to the output fName = os.path.split(file) data.name = fName[1] data.attrs['file'] = fName[1] data.attrs['fileBase'] = fName[0] print('Processed {0} sets of {1} file segments, ({2} blank)'.format(len(dumpSegs),recordType,blankSegs)) # Put in a list for now, might want to use Xarray dataset here, and/or combine results from multiple files. dataSet.append(data) return dataSet
# **************** Functions for wavefunction (3D data) files # Based on previous Matlab code, readOrb3D.m # Helped by Scipy Cookbook LAS reader example: https://scipy-cookbook.readthedocs.io/items/LASReader.html # # See: # - ePSproc_dev_IOfun_260519.py # - ePSproc_dev_3Dvis_290519.property # for development notes (Finn E:\ePS_paraview_proc\scripts). # # TODO: rename & integrate with functions above. # Read header lines into a list & convert to int
[docs]def readOrbHeader(f): headerLines=[] for i in range(5): headerLines.append(f.readline()) # headerLines.append(f.readline().split()) # Split at whitespace if i>0: headerLines[i] = int(headerLines[i]) # Convert to int return headerLines
# Read a specified number of floats from file, return as numpy array type
[docs]def readOrbElements(f,n): data = [] while len(data) < n: data.extend([float(s) for s in f.readline().split()]) return np.array(data)
# Read coords from file, based on params in headerLines
[docs]def readOrbCoords(f, headerLines): coords = [] for n in range(np.abs(headerLines[1])): coords.append(readOrbElements(f,headerLines[n+2])) return coords
# Read data from file, based on params in headerLines
[docs]def readOrbData(f, headerLines): # Determine size of grid nData = 1 nGrid = [] # [j=j*headerLines[k] for k in range(np.abs(headerLines[1]))] # FAIL # Set number of elements for n in range(np.abs(headerLines[1])): nData = nData*headerLines[n+2] nGrid.append(headerLines[n+2]) # Read elements dataRaw = readOrbElements(f,nData) # Resort to 3D array - adapted from Matlab code, so might be a +/-1 index offset mix-up here... C = 0 data = np.zeros(nGrid) for z in range(nGrid[2]): for y in range(nGrid[1]): # data[0:(nGrid[0]-1),y,z] = dataRaw[C:(C+nGrid[0]-1)] # Incorrect offset? data[:,y,z] = dataRaw[C:(C+nGrid[0])] # Should be correct... implicit (-1) in dataRaw range setting C = C+nGrid[0] return data
# *************** Master function for reading a set of 3D data files from ePS
[docs]def readOrb3D(fileIn = None, fileBase = None, fType = '_Orb.dat'): """ Read ePS 3D data file(s) and return results. File endings specified by fType, default *_Orb.dat. Parameters ---------- fileIn : str, list of strs, optional. File(s) to read (file in working dir, or full path). Defaults to current working dir if only a file name is supplied. For consistent results, pass raw strings, e.g. fileIn = r"C:\share\code\ePSproc\python_dev\no2_demo_ePS.out" fileBase : str, optional. Dir to scan for files. Currently only accepts a single dir. Defaults to current working dir if no other parameters are passed. fType : str, optional File ending for ePS output files, default '_Orb.dat' Returns ------- list List of data arrays, containing matrix elements etc. from each file scanned. # TODO: Change output to Xarray? Examples -------- >>> dataSet = readOrb3D() # Scan current dir >>> fileIn = r'C:\share\code\ePSproc\python_dev\DABCOSA2PPCA2PP_10.5eV_Orb.dat' >>> dataSet = readOrb3D(fileIn) # Scan single file >>> dataSet = readOrb3D(fileBase = r'C:\share\code\ePSproc\python_dev') # Scan dir """ # Populate file list fList = getFiles(fileIn = fileIn, fileBase = fileBase, fType = fType) dataSet = [] for fileName in fList: with open(fileName,'r') as f: # Check eof f.seek(0,2) fEnd = f.tell() f.seek(0) # Read file segments until eof # TODO: add eof checking here, can have 3 or 6 segments depending on symmetry. data = [] # for seg in range(3): while f.tell() < fEnd - 100: # Scan until eof minus arb small offset. headerLines = readOrbHeader(f) coords = readOrbCoords(f,headerLines) data.append(readOrbData(f, headerLines)) dataSet.append([fileName, headerLines, coords, data]) return dataSet
[docs]def writeOrb3Dvtk(dataSet): """ Write ePS 3D data file(s) to vtk format. This can be opened in, e.g., Paraview. Parameters ---------- dataSet : list List of data arrays, containing matrix elements etc. from each file scanned. Assumes format as output by readOrb3D(), [fileName, headerLines, coords, data] # TODO: Change to Xarray? Returns ------- list List of output files. .. note:: Uses Paulo Herrera's eVTK, see: * https://pyscience.wordpress.com/2014/09/06/numpy-to-vtk-converting-your-numpy-arrays-to-vtk-arrays-and-files/ * https://bitbucket.org/pauloh/pyevtk/src/default/ """ fOut = [] for file in dataSet: # Set grid, convert to Cart if necessary, assuming that grid won't be larger than 10 Angs if (len(file[2][0]) > 50): # Convert to Cart grid for plotting # TODO: Investigate use of sph grid here - should be cleaner. # TODO: Investigate recreating mesh in Paraview, rather than saving to file. [T,R,P] = np.meshgrid(file[2][1], file[2][0], file[2][2]) T = (T*np.pi/180) #-np.pi/2 P = P*np.pi/180 x = R*np.sin(P)*np.cos(T) z = R*np.cos(P) y = R*np.sin(P)*np.sin(T) else: [x,y,z] = np.meshgrid(file[2][1], file[2][0], file[2][2]) # Save single dataset # Function info: https://bitbucket.org/pauloh/pyevtk/src/default/src/hl.py # gridToVTK("./ePStest3", x, y, z, pointData = {'data': data[0]}) # Save full dataset # TODO: Check number of segments and save multiple # segs = round(len(file[3])/3) fOut.append(gridToVTK(file[0][:-4], x, y, z, pointData = {'Re': file[3][0], 'Im': file[3][1], 'Abs': file[3][2]})) print("{0} files written to vtk format.".format(len(fOut))) return fOut
#**************** Wrappers for Xarray load/save netCDF # File write wrapper.
[docs]def writeXarray(dataIn, fileName = None, filePath = None): """ Write file to netCDF format via Xarray method. Parameters ----------- dataIn : Xarray Data array to write to disk. fileName : str, optional, default = None Filename to use. If set to None (default) the file will be written with a datastamp. filePath : str, optional, default = None Full path to file. If set to None (default) the file will be written in the current working directory (as returned by `os.getcwd()`). Returns ------- str Indicates save type and file path. Notes ----- The default option for Xarray is to use Scipy netCDF writer, which does not support complex datatypes. In this case, the data array is written as a dataset with a real and imag component. TODO: implement try/except to handle various cases here, and test other netCDF writers (see http://xarray.pydata.org/en/stable/io.html#netcdf). Multi-level indexing is also not supported, and must be serialized first. Ugh. """ if fileName is None: timeString = dt.now() fileName = 'ep_' + timeString.strftime('%Y-%m-%d_%H-%M-%S') if filePath is None: filePath = os.getcwd() # Serialize MultiIndex - testing here for BLM case. # if 'BLM' in dataIn.dims: # dataIn = dataIn.reset_index(['Euler','BLM']) # Serialize general - use unstact() to flatten all dims dataIn = dataIn.unstack() # Try/except not yet working, multiple error types to handle here... # try: # dataIn.to_netcdf(fileName) # saveMsg = 'Written to netCDF4.' # print(saveMsg) # return saveMsg # # except ValueError as e: # if e.msg != "NetCDF 3 does not support type complex128": # raise # else: # xr.Dataset({'Re':dataIn.real, 'Im':dataIn.imag}).to_netcdf(fileName) # saveMsg = 'Written to netCDF3 (re/im format).' # print(saveMsg) # return saveMsg # # return 'File not written.' # Safe version with re/im split save type only. # xr.Dataset({'Re':dataIn.real, 'Im':dataIn.imag}).to_netcdf(os.path.join(filePath, fileName + '.nc')) # This works, but drops attrs! dataOut = xr.Dataset({'Re':dataIn.real, 'Im':dataIn.imag}) dataOut.attrs = dataIn.attrs # Allow for SF & XS coords which may also be complex if 'XS' in dataOut.coords: dataOut['XSr'] = dataOut.XS.real dataOut['XSi'] = dataOut.XS.imag dataOut = dataOut.drop('XS') if 'SF' in dataOut.coords: dataOut['SFr'] = dataOut.SF.real dataOut['SFi'] = dataOut.SF.imag dataOut = dataOut.drop('SF') dataOut.to_netcdf(os.path.join(filePath, fileName + '.nc')) saveMsg = ['Written to netCDF3 (re/im format)'] saveMsg.append(os.path.join(filePath, fileName + '.nc')) print(saveMsg) return saveMsg
# File read wrapper.
[docs]def readXarray(fileName, filePath = None): """ Read file from netCDF format via Xarray method. Parameters ----------- fileName : str File to read. filePath : str, optional, default = None Full path to file. If set to None (default) the file will be written in the current working directory (as returned by `os.getcwd()`). Returns ------- Xarray Data from file. May be in serialized format. Notes ----- The default option for Xarray is to use Scipy netCDF writer, which does not support complex datatypes. In this case, the data array is written as a dataset with a real and imag component. Multi-level indexing is also not supported, and must be serialized first. Ugh. TODO: generalize multi-level indexing here. """ # Read file dataIn = xr.open_dataset(fileName) # Reconstruct complex variables, NOTE this drops attrs... there's likely a better way to do this! dataOut = dataIn.Re + dataIn.Im*1j dataOut.attrs = dataIn.attrs # Rest SF & XS coords which may also be complex # Note: need to check vs. dataIn here, since dataOut already has dropped vars if 'XSr' in dataIn.data_vars: dataOut['XS'] = dataIn.XSr + dataIn.XSi*1j # dataOut = dataOut.drop('XSr').drop('XSi') if 'SFr' in dataIn.data_vars: dataOut['SF'] = dataIn.SFr + dataIn.SFi # dataOut = dataOut.drop('SFr').drop('SFi') # Recreate MultiIndex from serialized version - testing here for BLM case. # if 'BLM' in dataIn.dims: # dataIn = dataIn.set_index({'BLM':['l','m'],'Euler':['P','T','C']}) # Recreate MultiIndex from serialized version according to array type. if dataIn.dataType == 'BLM': dataOut = dataOut.stack(BLMdimList(sType = 'sDict')) elif dataIn.dataType == 'matE': dataOut = dataOut.stack(matEdimList(sType = 'sDict')) return dataOut