3.7.1. ADF Suite¶
PLAMS offers interfaces to three main binaries of the ADF Suite: ADF, BAND and DFTB as well as some other small utility binaries like DENSF of FCF. All possible input keywords and options are covered, as well as extraction of arbitrary data from binary files (called KF files) produced by these programs.
3.7.1.1. ADF, BAND and DFTB¶
ADF, BAND and DFTB are of course very different programs, but from our perspective they are rather similar. Their input files follow a common structure of blocks and subblocks. They store results as binary files in KF format and and they print human-readable summary of calculation to the standard output. They also share command line arguments, error messages etc. Thanks to that Python code responsible for creating, running and examining jobs with ADF, BAND and DFTB jobs overlaps a lot and can be grouped together in abstract classes. SCMJob and SCMResults are subclasses of, respectively, SingleJob and Results and serve as bases for concrete classes: ADFJob, BANDJob, DFTBJob, ADFResults, BANDResults and DFTBResults. Code contained in these concrete classes describes small technical differences and is used only internally, so they are omitted in the API specification below. From user perspective they all follow the common interface defined by SCMJob and SCMResults. That means in your scripts you would create instances of ADFJob, BANDJob or DFTBJob, but methods that you can use with them (and their corresponding results) can be taken from SCMJob and SCMResults.
3.7.1.1.1. Preparing input¶
Although input files for ADF, BAND and DFTB use different sets of keywords, they all have the same logical structure – they consist of blocks and subblocks containg keys and values. That kind of structure can be easily reflected by Settings objects since they are built in a similar way.
The input file is generated based on input branch of job’s Settings. All data present there is translated to input contents. Nested Settings instances define blocks and subblocks, as in the example below:
>>> myjob = ADFJob(molecule=Molecule('water.xyz'))
>>> myjob.settings.input.basis.type = 'DZP'
>>> myjob.settings.input.basis.core = 'None'
>>> myjob.settings.input.basis.createoutput = 'None'
>>> myjob.settings.input.scf.iterations = 100
>>> myjob.settings.input.scf.converge = '1.0e-06 1.0e-06'
>>> myjob.settings.input.save = 'TAPE13'
Input file created during execution of myjob looks like:
Atoms
#coordinates from water.xyz
End
Basis
Createoutput None
Core None
Type DZP
End
Save TAPE13
Scf
Converge 1.0e-06 1.0e-06
Iterations 100
End
As you can see, entries present in myjob.settings.input. are listed in the alphabetical order. If an entry is a regular key-value pair it is printed in one line (like Save TAPE13 above). If an entry is a nested Settings instance it is printed as a block and entries in this instance correspond to contents of a the block. All keys inside Settings are lowercased and the first letter is later capitalized when printing the input file. Values on the other hand remain unchanged. Strings put as values can contain spaces like converge above – the whole string is printed after the key. That allows to handle lines that need to contain more than one key=value pair. If you need to put a key without any value, True or empty string can be given as a value:
>>> myjob.settings.input.geometry.SP = True
>>> myjob.settings.input.writefock = ''
# translates to:
Geometry
Sp
End
Writefock
To produce an empty block simply type:
>>> myjob.settings.input.geometry # this is equivalent to myjob.settings.input.geometry = Settings()
#
Geometry
End
The algorithm translating Settings contents into input file does not check the correctness of the data - it simply takes keys and values from Settings instance and puts them in the text file. Due to that you are not going to be warned if you make a typo, use wrong keyword or improper syntax. Beware of that.
>>> myjob.settings.input.dog.cat.apple = 'pear'
#
Dog
Cat
Apple pear
Subend
End
Some blocks require (or allow) some data to be put in the header line, next to the block name. Special key _h is helpful in these situations:
>>> myjob.settings.input.someblock._h = 'header=very important'
>>> myjob.settings.input.someblock.key1 = 'value1'
>>> myjob.settings.input.someblock.key2 = 'value2'
#
Someblock header=very important
Key1 value1
Key2 value2
End
The order of blocks within input file and subblocks within a parent block follows Settings iteration order which is lexicographical (however, SCMJob is smart enough to put blocks like DEFINE or UNITS at the top of the input). In rare cases you would want to override this order, for example when you supply ATOMS block manually, which can be done when automatic molecule handling is disabled (see below). That behavior can be achieved by another type of special key:
>>> myjob.settings.input.block._1 = 'entire line that has to be the first line of block'
>>> myjob.settings.input.block._2 = 'second line'
>>> myjob.settings.input.block._4 = 'I will not be printed'
>>> myjob.settings.input.block.key1 = 'value1'
>>> myjob.settings.input.block.key2 = 'value2'
#
Block
entire line that has to be the first line of block
second line
Key1 value1
Key2 value2
End
Sometimes one needs to put more instances of the same key within one block, like for example in CONSTRAINTS block in ADF. It can be done by using list of values instead of a single value:
>>> myjob.settings.input.constraints.atom = [1,5,4]
>>> myjob.settings.input.constraints.block = ['ligand', 'residue']
#
Constraints
Atom 1
Atom 5
Atom 4
Block ligand
Block residue
End
Finally, in some rare cases key and value pair in the input needs to be printed in a form key=value instead of key value. When value is a string starting with the equal sign, no space is inserted between key and value:
>>> myjob.settings.input.block.key = '=value'
#
Block
Key=value
End
Sometimes a value of a key in the input file needs to be a path to some file, usually KF file with results of some previous calculation. Of course such a path can be given explicitly newjob.restart = '/home/user/science/plams.12345/oldjob/oldjob.t21', but for user’s convenience instances of SCMJob or SCMResults (or directly KFFile) can be also used. Algorithm will detect it and use an absolute path to the main KF file instead:
>>> myjob.settings.input.restart = oldjob
>>> myjob.settings.input.fragment.frag1 = fragjob
#
Restart /home/user/science/plams.12345/oldjob/oldjob.t21
Fragment
Frag1 /home/user/science/fragmentresults/somejob/somejob.t21
End
Molecule instance stored in job’s molecule attribute is automatically processed during the input file preparation and printed in the proper format, depending on the program. It is possible to disable that and give molecular coordinates explicitly as entries in myjob.settings.input.. Automatic molecule processing can be turned off by myjob.settings.ignore_molecule = True.
3.7.1.1.2. Special atoms in ADF¶
- In ADF atomic coordinates in
ATOMSblock can be enriched with some additional information like special names of atoms (for example in case of using different isotopes) or block/fragment membership. Since usually contents ofATOMSblock are generated automatically based on theMoleculeassociated with a job, this information needs to be supplied inside the givenMoleculeinstance. Details about every atom can be adjusted separately, by modifying attributes of a particularAtominstance according to the following convention: - Atomic symbol is generated based on atomic number stored in
atnumattribute of a correspondingAtom. Atomic number 0 corresponds to the “dummy atom” for which the symbol is empty. - If an attribute
ghostof anAtomisTrue, the above atomic symbol is prefixed withGh.. - If an
Atomhas an attributenameits contents are added after the symbol. Hence settingatnumto 0 and adjustingnameallows to put an arbitrary string as the atomic symbol. - If an
Atomhas an attributefragmentits contents are added after atomic coordinates withf=prefix. - If an
Atomhas an attributeblockits contents are added after atomic coordinates withb=prefix.
- Atomic symbol is generated based on atomic number stored in
The following example illustrates the usage of this mechanism:
>>> mol = Molecule('xyz/Ethanol.xyz')
>>> mol[0].ghost = True
>>> mol[1].name = 'D'
>>> mol[2].ghost = True
>>> mol[2].name = 'T'
>>> mol[3].atnum = 0
>>> mol[3].name = 'J.XYZ'
>>> mol[4].atnum = 0
>>> mol[4].name = 'J.ASD'
>>> mol[4].ghost = True
>>> mol[5].fragment = 'myfragment'
>>> mol[6].block = 'block1'
>>> mol[7].fragment = 'frag'
>>> mol[7].block = 'block2'
>>> myjob = ADFJob(molecule=mol)
#
Atoms
1 Gh.C 0.01247 0.02254 1.08262
2 C.D -0.00894 -0.01624 -0.43421
3 Gh.H.T -0.49334 0.93505 1.44716
4 J.XYZ 1.05522 0.04512 1.44808
5 Gh.J.ASD -0.64695 -1.12346 2.54219
6 H 0.50112 -0.91640 -0.80440 f=myfragment
7 H 0.49999 0.86726 -0.84481 b=block1
8 H -1.04310 -0.02739 -0.80544 f=frag b=block2
9 O -0.66442 -1.15471 1.56909
End
3.7.1.1.3. Preparing runscript¶
Runscripts for ADF, BAND and DFTB are very simple - they are just single execution of one of the binaries with proper standard input and output handling. The number of parallel processes (-n parameter) can be adjusted with myjob.settings.runscript.nproc.
3.7.1.1.4. Results extraction¶
All three programs print results to the standard output. The output file can be examined with standard text processing tools (grep_output() and awk_output()). Besides that all calculation details are saved in the binary file in KF format. This file is called TAPE21 for ADF, RUNKF for BAND and dftb.rkf for DFTB. PLAMS renames those files to, respectively [jobname].t21, [jobname].runkf and [jobname].rkf. Data stored in those files can be accessed using additional methods defined in SCMResults class.
3.7.1.1.5. API¶
-
class
SCMJob(molecule=None, name='plamsjob', settings=None, depend=None)[source]¶ Abstract class gathering common mechanisms for jobs with all ADF Suite binaries.
-
get_input()[source]¶ Transform all contents of
setting.inputbranch into string with blocks, keys and values.On the highest level alphabetic order of iteration is modified: keys occuring in attribute
_topare printed first.Automatic handling of
moleculecan be disabled withsettings.ignore_molecule = True.
-
get_runscript()[source]¶ Generate a runscript. Returned string is of the form:
$ADFBIN/name [-n nproc] <jobname.in [>jobname.out]
nameis taken from the class attribute_command.-nflag is added ifsettings.runscript.nprocexists.[>jobname.out]is used based onsettings.runscript.stdout_redirect.
-
check()[source]¶ Check if
termination statusvariable fromGeneralsection of main KF file equalsNORMAL TERMINATION.
-
_parsemol()[source]¶ Process
Moleculeinstance stored inmoleculeattribute and add it as relevant entries ofsettings.inputbranch. Abstract method.
-
_removemol()[source]¶ Remove from
settings.inputall entries added by_parsemol(). Abstract method.
-
_settings_reduce()[source]¶ When this object is present as a value in some
Settingsinstance and string representation is needed, use the absolute path to the main KF file. SeeSettings.__reduce__for details.
-
-
class
SCMResults(job)[source]¶ Abstract class gathering common mechanisms for results of all ADF Suite binaries.
-
collect()[source]¶ Collect files present in the job folder.
Use parent method from
Results, then create an instance ofKFFilefor the main KF file and store it as_kfattribute.
-
readkf(section, variable)[source]¶ Read data from section/variable of the main KF file.
The type of the returned value depends on the type of variable defined inside KF file. It can be: single int, list of ints, single float, list of floats, single boolean, list of booleans or string.
-
newkf(filename)[source]¶ Create new
KFFileinstance using file filename in the job folder.Example usage:
>>> res = someadfjob.run() >>> tape13 = res.newkf('$JN.t13') >>> print(tape13.read('Geometry', 'xyz'))
-
get_molecule(section, variable, unit='bohr', internal=False, n=1)[source]¶ Read molecule coordinates from section/variable of the main KF file.
Returned
Moleculeinstance is created by copying a molecule from associatedSCMJobinstance and updating atomic coordinates with values read from section/variable. The format in which coordinates are stored is not consistent for all programs or even for different sections of the same KF file. Sometimes coordinates are stored in bohr, sometimes in angstrom. The order of atoms can be either input order or internal order. These settings can be adjusted with unit and internal parameters. Some variables store more than one geometry, in those cases n can be used to choose the preferred one.
-
_settings_reduce()[source]¶ When this object is present as a value in some
Settingsinstance and string representation is needed, use the absolute path to the main KF file. SeeSettings.__reduce__for details.
-
_export_attribute(attr, other)[source]¶ If attr is a KF file take care of a proper path. Otherwise use parent method. See
Results._copy_tofor details.
-
3.7.1.2. Other tools: Densf, FCF¶
Apart from main computational programs mentioned above, ADFSuite offers a range of small utility tools that can be used to obtain more specific results. Those tools usually base on the prior run of one of the main programs and need the KF file produced by them as a part of the input.
From the functional point of view these tools are very similar to ADF, BAND and DFTB. Their results are stored in KF files and their input files follow the same structure of blocks, keys and values. Because of that the same classes (SCMJob and SCMResults) are used as bases and hence preparation, running and results extraction for utility tools follow the rules described above, in ADF, BAND and DFTB
The main difference is that usually utility jobs don’t need molecular coordinates as part of the input (they extract this information from previous calculation’s KF file). So no Molecule instance is needed and the molecule attribute of the job object is simply ignored. Because of that get_molecule() method does not work with FCFResults, DensfResults etc.
Below you can find the list of dedicated job classes that are currently available. Details about input specification for those jobs can be found in corresponding part of ADF Suite documentation.
-
class
FCFJob(inputjob1=None, inputjob2=None, name='plamsjob', settings=None, depend=None)[source]¶ A class representing calculation of Franck-Condon factors using
fcfprogram.Two new attributes are introduced:
inputjob1andinputjob2. They are used to supply KF files from previous runs tofcfprogram. The value can either be a string with a path to KF file or an instance of any type ofSCMJoborSCMResults(in this case the path to corresponding KF file will be extracted automatically). If the value ofinputjob1orinputjob2isNone, no automatic handling occurs and user needs to manually supply paths to input jobs using proper keywords placed inmyjob.settings.input(STATESorSTATE1andSTATE2).The resulting
TAPE61file is renamed tojobname.t61.
-
class
DensfJob(inputjob=None, name='plamsjob', settings=None, depend=None)[source]¶ A class representing calculation of molecular properties on a grid using
densfprogram.A new attribute
inputjobis introduced to supply KF file from previously run job. The value can either be a string with a path to KF file or an instance of any type ofSCMJoborSCMResults(in this case the path to corresponding KF file will be extracted automatically). If the value ofinputjobisNone, no automatic handling occurs and user needs to manually supply path to input job usingINPUTFILEkeyword placed inmyjob.settings.input.The resulting
TAPE41file is renamed tojobname.t41.
3.7.1.3. KF files¶
KF is the main format for storing binary data used in all ADFSuite programs. PLAMS offers an easy and efficient way of accessing the data stored in existing KF files, as well as modifying and creating them.
3.7.1.3.1. KFFile¶
-
class
KFFile(path, autosave=True)[source]¶ A class for reading and writing binary files in KF format.
This class acts as a wrapper around
KFReadercollecting all the data written by user in some “temporary zone” and using Fortran binariesudmpkfandcpkfto write this data to the physical file when needed.The constructor argument path should be a string with a path to an existing KF file or a new KF file that you wish to create. If a path to existing file is passed, new
KFReaderinstance is created allowing to read all the data from this file.When
write()method is used, the new data is not immediately written to a disk. Instead of that, it is temporarily stored intmpdatadictionary. When methodsave()is invoked, contents of that dictionary are written to a physical file andtmpdatais emptied.Other methods like
read()ordelete_section()are aware oftmpdataand work flawlessly, regardless ifsave()was called or not.By default,
save()is automatically invoked after eachwrite(), so physical file on a disk is always “actual”. This behavior can be adjusted with autosave constructor parameter. Having autosave enabled is usually a good idea, however, if you need to write a lot of small pieces of data to your file, the overhead of callingudmpkfandcpkfafter everywrite()can lead to significant delays. In such a case it is advised to disable autosave and callsave()manually, when needed.Dictionary-like bracket notation can be used as a shortcut to read and write variables:
mykf = KFFile('someexistingkffile.kf') #all three below are equivalent x = mykf['General%Termination Status'] x = mykf[('General','Termination Status')] x = mykf.read('General','Termination Status') #all three below are equivalent mykf['Geometry%xyz'] = somevariable mykf[('Geometry','xyz')] = somevariable mykf.write('Geometry','xyz', somevariable)
-
read(section, variable)[source]¶ Extract and return data for a variable located in a section.
For single-value numerical or boolean variables returned value is a single number or bool. For longer variables this method returns a list of values. For string variables a single string is returned.
-
write(section, variable, value)[source]¶ Write a variable with a value in a section . If such a variable already exists in this section, the old value is overwritten.
-
__getitem__(name)[source]¶ Allow to use
x = mykf['section%variable']orx = mykf[('section','variable')]instead ofx = kf.read('section', 'variable').
-
__setitem__(name, value)[source]¶ Allow to use
mykf['section%variable'] = valueormykf[('section','variable')] = valueinstead ofkf.write('section', 'variable', value).
-
_settings_reduce()[source]¶ When this object is present as a value in some
Settingsinstance and string representation is needed, use the absolute path. SeeSettings.__reduce__for details.
-
3.7.1.3.2. KFReader¶
-
class
KFReader(path, blocksize=4096, autodetect=True)[source]¶ A class for efficient Python-native reader of binary files in KF format.
This class offers read-only access to any fragment of data from a KF file. Unlike other Python KF readers, this one does not use the Fortran binary
dmpkfto process KF files, but instead reads and interprets raw binary data straight from the file, on Python level. That approach results in significant speedup (by a factor of few hundreds for large files extracted variable by variable).The constructor argument path should be a string with a path (relative or absolute) to an existing KF file.
blocksize indicates the length of basic KF file block. So far, all KF files produced by any of ADFSuite programs have the same block size of 4096 bytes. Unless you’re doing something very special, you should not touch this value.
Organization of data inside KF file can depend on a machine on which this file was produced. Two parameters can vary: the length of integer (32 or 64 bit) and endian (little or big). These parameters have to be determined before any reading can take place, otherwise the results will have no sense. If the constructor argument autodetect is
True, the constructor attempts to automatically detect the format of a given KF file, allowing to read files created on a machine with different endian or integer length. This automatic detection is enabled by default and it is advised to leave it that way. If you wish to disable it, you should setendianandwordattributes manually before reading anything (see the code for details).Note
This class consists of quite technical, low level code. If you don’t need to modify or extend
KFReader, you can safely ignore all private methods, all you need isread()and occasionally__iter__()-
read(section, variable)[source]¶ Extract and return data for a variable located in a section.
For single-value numerical or boolean variables returned value is a single number or bool. For longer variables this method returns a list of values. For string variables a single string is returned.
-
_settings_reduce()[source]¶ When this object is present as a value in some
Settingsinstance and string representation is needed, use the absolute path. SeeSettings.__reduce__for details.
-
_autodetect()[source]¶ Try to automatically detect the format (int size and endian) of this KF file.
-
_parse(block, format)[source]¶ Translate a block of binary data into list of values in specified format.
format should be a list of pairs (a,t) where t is one of the following characters:
's'for string,'i'for 32-bit integer,'q'for 64-bit integer and a is the number of occurrences (or length of a string).For example, if format is equal to
[(32,'s'),(4,'i'),(2,'d'),(2,'i')], the contents of block are divided into 72 bytes (32*1 + 4*4 + 2*8 + 2*4 = 72) chunks (possibly droping the last one, if it’s shorter than 72 bytes). Then each chunk is translated to a 9-tuple of string, 4 ints, 2 floats and 2 ints. List of such tuples is the returned value.
-
_get_data(datablock)[source]¶ Extract all data from a single data block. Returned value is a 4-tuple of lists, on list for each data type (respectively: int, float, str, bool).
-
_create_index()[source]¶ Find and parse relevant index blocks of KFFile to extract the information about location of all sections and variables.
Two dictionaries are populated during this process.
_datacontains, for each section, a list of triples describing how logical blocks of data are mapped into physical ones. For example,_data['General'] = [(3,6,12), (9,40,45)]means that logical blocks 3-8 of sectionGeneralare located in physical blocks 6-11 and logical blocks 9-13 in physical blocks 40-44. This list is always sorted via first tuple elements allowing efficient access to arbitrary logical block of each section.The second dictionary,
_sections, is used to locate each variable within its section. For each section, it contains another dictionary of each variable of this section. So_section[sec][var]contains all information needed to extract variablevarfrom sectionsec. This is a 4-tuple containing the following information: variable type, logic block in which the variable first occurs, position within this block where its data start and the length of the variable. Combining this information with mapping stored in_dataallows to extract each single variable.
-