|Version 20 (modified by 3 years ago) ( diff ),|
Table of Contents
- Table Creation
- Duplicate Entries
- Constant value columns
- In case of failure
This page describes two tools which are - apart from their input file format - nearly identical:
Both file types, ROOT and FITS files, have storage engines for the storage of table like data. In ROOT files these are trees with branches and leaves and in FITS files these are FITS tables, the so-called extensions. Although, root offers a lot of printing and fitting options, sometimes the access through a database has advantages. Therefore, a tool which maps ROOT-files and FITS-files into SQL tables helps.
First get the man page of
root2sql either with
root2sql --help or (if installed) with
man root2sql. This page will mainly outline how
root2sql works, because
fits2sql is slightly less complex. In short, the differences will be explained at the end of the document.
How a connection to the database is established is outlined in: DatabaseBasedAnalysis/Connection. Per default, your connection will be compressed if the the host-address of the mysql-server is neither localhost nor 127.0.0.1. To force compression use
Generally, compression should be enable if connecting from the world. To test compression, the "Hello World! " example can be used. If you are using an ssh tunnel, make sure that you do not compress your stream twice!
In the following it is assumed that the database connection is working and tested, so that the resources (
uri=...) are omitted.
Generally, ROOT-files can store very complicated object as a TH1F. Although, they could be entered in a database as a so called BLOB (a binary object), this is not the idea of the tool. Thus only basic data types are supported.
The input file is given as the first positional argument
root2sql [--file] inputfile.root
Each root tree has branches and leaves (the basic data types). These leaves can be read independently of the classes which were used to write the root file. The default tree to read from is Events but the name can be overwritten using
-t (an abbreviation of
--tree). Which would look like this
root2sql inputfile.root -t MyTree
The default table which is filled with the data in this tree is expected to have the same name than the tree. If this is not the case, it can be changed with
root2sql inputfile.root -t MyTree --table MyTable
In these three examples, the first one would read from a tree Events and write to table Events. The second case would read from a tree MyTree and write to a table MyTree and the third case would read from a tree MyTree and write to a table MyTable.
The name of each column to which data is filled from a leave is obtained from the leaves' names. The leave names can be checked using
--print-branches exists for convenience to print only the high-level branches. If you have not yet configured database access you can work with an empty uri
If the leaf with the basic data type is called MTime.fTime.fMilliSec, the column would be called MTime.fTime.fMilliSec as well.
Both, root leaves and FITS-extension can contain fixes sized arrays of basic data types. If the array is called ARR and has 3 fields, the SQL columns will be called ARR, ARR and ARR. All modifier options working on columns always alter all corresponding columns at once.
Changing Column Names
Sometimes the leaves' names might be quite inconvenient like in the example above. To allow to simplify column names, regular expressions (using boost's regex) can be defined to change the names. A helpful resource for regex testing is https://www.regextester.com/. Note, that these regular expressions are applied one by one on each leaf's name. A valid expression could be:
which would remove all occurrences of MHillas.f.
would replace all occurrences of MHillas.f by MHillas_.
The corresponding call in C++ is
boost::regex_replace(leafname, search, replace) for
--map option can be given more than once, many and complex replacement rules can be defined.
They are applied in sequence, one by one and to all leaves. A single match does not stop the sequence.
Sometimes it might also be convenient to skip a leaf. This can be done with the
--ignore regex option. If the given regular expression yields a match, the leaf will be ignored.
Note that the regular expression works on the raw-name of the leaf not the readily mapped SQL column names.
will skip all leaves which name contains with 'ThetaSq.'. This option can be given more than once.
The corresponding C++ call is
boost::regex_match(leafname, regex) for
Column Data Types
The data type of each column is kept as close as possible to the leaves' basic data types. If for some reason this is not wanted, the data type of the SQL column can be overwritten with
--sql-type column/type. This works like the following:
In this example, FileId is the name of the SQL column for which the data type UNSIGNED INT should be used. The second column (UNSIGNED INT) is a basic SQL data type. The option can be given more than once.
It is not very handy if the SQL table has to be created manually beforehand. Therefore,
root2sql can automatically create the table. To do this, specify
--create. The query which would be used to create the table can be printed with
--print-create even when
--create has not been specified. Tables are created only if the do not yet exist (
CREATE TABLE IF NOT EXISTS). To ensure recreation of a table, the old table can be dropped with
--drop. As any insert operation would fail if the required table is dropped and no new one is created,
--drop always implies
All columns of the table are created as
NOT NULL as default.
The default database engine is to create the table with the default of the database. If a database engine (
ENGINE=) should be forced for the table the
--engine ENGINE option can be used. Similarly, the row-format (
ROW_FORMAT=) can be defined using
--row-format FORMAT. For example:
--engine InnoDB --row-format=COMPRESSED
For convenience, a primary key can be defined by the the columns which belong to the primary key.
To choose the columns which should become primary keys, use
--primary, for example:
Note that the combination of these columns must be unique.
INSERT query would fail if another row with the same primary key exists already. This can be avoided adding the
IGNORE keyword to the
INSERT query by
--ignore-errors, which essentially ignores all errors and turns them into warnings which are printed after the query succeede.
Another option is to update duplicate entries. This can be done using the
ON DUPLICATE KEY UPDATE directive. Giving
-duplicate, you can specify what should be updated in case of a duplicate key. To keep the row untouched, you can just update the primary key with the identical primary key, e.g.
--duplicate resource can be specified more than once to add more expressions to the assignment_list. For more details, see the MySQL manual.
Constant value columns
Add a constant value
Sometimes it might be necessary to insert a constant value for each file which is processed (for example a file ID which is not part of the stored data). This can be achieved by the
--fixed.COLUMN=VALUE directive where COLUMN the column name and VALUE is the value filled into the column. As default, the column is create as INT UNSIGNED but this can be overwritten as usual with the
A special case is when VALUE has the form
/regex/[format]/. In this case the filename is replaced given the regular expressen by the format using boost::regex_replace. Replacement uses the standard formatting rules to replace matches (those used by ECMAScript's replace method). If the format is omitted, the default is
$0 which corresponds to just taking the match. For example, if
would be match the _1234_ in the filename file_12345_name.ext. It would be replaced by the first sub-sequence (%1) which is '12345'. ($0 is the whole match _12345_).
Based on the constant values, conditional execution can be enable with
--conditional. After dropping/creating the table, it is checked by a
SELECT query whether at least one entry which fits the constant value columns exists. If at least one exists, further execution is stopped.
Usually the previously defined constant values are helpful to create an index which relates unambiguously the inserted data to the file. It might be useful to delete all data which belongs to this particular file before new data is entered. This can be achieved with the
--delete directive. It deletes all data from the table before inserting new data which fulfills the condition defined by the
To do the pre-deleting efficiently, it makes sense to have an INDEX created for a fast access. This can be achieved with the
--index directive which adds the corresponding statement to the table creation (
For debugging purpose, or to just create or drop a table, the final insert query can be skipped using
--no-insert. All operation which would alter anything in the database are skipped using
--dry-run (note that this might create consequent errors).
Additional information about the connection are printed using
--print-connection. Note that this implies additional queries and more traffic and is therefore slightly slower.
Using a higher verbosity level (-v), an overview of the written columns or all processed leaves is printed depending on the verbosity level. The output looks like the following
Leaf name [root data type] (SQL name)
MTime.fTime.fMilliSec [Long64_t] (MilliSec)
which means that the leaf MTime.fTime.fMilliSec is detected to be a Long64_t which is filled into a column called MilliSec. Leaves with non basic data types are ignored automatically and are marked as (-n/a-). User ignored columns are marked as (-ignored-).
INSERT query can be printed with
--print-insert. Note that this can be a quite long output.
For test purposes, the number of events and the events which are processed can be altered with
Note that for performance reason, all data is collected in memory and a single INSERT query is issued at the end.
In case of failure
If a query failed, the query is printed to stderr together with the error message. For the main INSERT query, this is only true if the verbosity level is at least 2 or the query has less than 80*25 bytes.
In case of success, 0 is returned, a value>0 otherwise.
fits2sql executable works mainly identical to the
root2sql executable, except that the ROOT-tree here is a so-called FITS-extension (table). Therefore, the
--tree option is called
--print-extension shows the columns of a FITS extension and
--print-columns a list of all columns.
As a default, integer columns in FITS-files contain signed values. To convert them to unsigned columns, the
--unsigned option can be given follows by a column name. It can be given more than once to apply that to more than one column.
For example, my resource file to convert ganymed files to SQL tables looks like this:
# Database uri and credentials #uri=... # Regular expression to match columns which shell be ignored ignore=ThetaSquared\..* ignore=Ghostbuster\..* ignore=MHillasSrc\..* ignore=MHillasSrcAnti\..* ignore=MSrcPosCam\..* ignore=MSrcPosAnti\..* ignore=DataType\..* ignore=.*\.fUniqueID ignore=.*\.fBits ignore=MPointingPos\.fHa # Regular expressions to simplify column names in table map=\.fVal/ map=MHillas\.f/ map=MHillasExt\.f/ map=MPointingPos\.f/ map=MImagePar\.f/ map=MNewImagePar\.f/ map=MTime\.fTime\.f/ map=MTime\.f/ # Overwrites the SQL data type for some columns sql-type=FileId/INT UNSIGNED sql-type=EvtNumber/INT UNSIGNED # Columsn to be setup as primary keys primary=FileId primary=EvtNumber