Professional Documents
Culture Documents
Clio Grows Up
Clio Grows Up
Correspondences GUI
Mapping
• XML Schema Generation • XML Schema
• Relational • Relational
Query
“conforms to” Generation
“conforms to”
SQL XQuery … XSLT
data
Executable transformation
(SQL/XQuery/XSLT/…)
S2 = ExpFactor ExpSet 2. ∃ t0 ∈ /GEML/exp_set, t1 ∈ t0/exp_set_header/exp_factor_list/exp_factor
S3 = TreatmentLevel ExpSet 3. such that s1/LOCAL_ACCESSION = t0 /@local_accession ∧
S4 = TL_FactorValue TreatmentLevel ⋈ ExpSet 4. s1/NAME = t0/@name ∧ s1/ES_PK = t0/@id ∧
ExpFactor ExpSet 5. s1/RELEASE_DATE = t0/@release_date ∧
6. s1/ANALYSIS_DESC = t0/exp_set_header/analysis_desc ∧
T1 = GEML/exp_set …
T2 = GEML/exp_set/…/exp_factor 1. m3: ∀ s0 ∈ /TL_FactorValue, s1 ∈ /TreatmentLevel, s2 ∈ /ExpSet,
T3 = GEML/exp_set/…/treatment 2. s3 ∈ /ExpFactor, s4 ∈ /ExpSet,
3. where s0/tl_fk = s1/tl_pk ∧ s1/es_fk = s2/es_pk ∧
T4 = GEML/exp_set/ { …/treatment/…/treat_factor ⋈ …/exp_factor }
4. s0/ef_fk = s3/ef_pk ∧ s3/es_fk = s4/es_pk
5. ∃ t0 ∈ /GEML/exp_set, t1 ∈ t0/exp_set_header/treatment_list/treatment,
Figure 3: Source and target tableaux (informal notation). 6. t2 ∈ t1/treat_factor_list/treat_factor, t3 ∈ t0/exp_set_header/exp_factor_list/exp_factor
7. where t2/factor_id = t3/id
8. such that …
the other set-type elements that can be reached by following foreign
key (keyref) constraints (a process called the chase). The result is a
set of tableaux2 , one set in each schema. Figure 4: Two logical mappings.
In Figure 3 we show several of the source and target tableaux maps easily to efficient iteration patterns (e.g., the from clause of
that are generated for our example. (For brevity, we sometimes use SQL, or the for loops of XQuery/XSLT).
ExpSet instead of EXPERIMENTSET; similar abbreviations are also Generation of logical mappings The second step of the mapping
used for the other tables.) For the source schema, ExpSet forms a generation algorithm is the generation of logical mappings. (Recall
tableau by itself, because ExpSet is a top-level table (there are no that a schema mapping is a set of logical mappings.) The basic algo-
outgoing foreign keys). In contrast, ExpFactor does not form a rithm in [7] pairs all the existing tableaux in the source with all the
tableau by itself but needs ExpSet into which it has a foreign key. A existing tableaux in the target, and then finds the correspondences
more complicated tableau is S4 that is constructed for TL FactorVa- that are covered by each pair. If there are such correspondences, then
lue. Each factor value is associated to one treatment level (thus, the the given pair of tableaux is a candidate of a logical mapping. (In
foreign key into TreatmentLevel) and each treatment level cor- Section 3 we describe an additional filtering that takes place before
responds to one experiment set (thus, the foreign key into ExpSet). such a candidate is actually output to a user.)
However, a factor value is also an experiment factor (thus, the foreign In Figure 4 we show two of the resulting logical mappings, for our
key into ExpFactor), and each experiment factor is associated to example. The mapping m1 is obtained from the pair (S2 , T2 ), which
an experiment set (hence, the second occurrence of ExpSet). is covered by 10 correspondences. The source tableau is encoded in
The above tableau S4 illustrates the complexity that can arise even the ∀ clause (with its associated where clause that stores the condi-
when a relatively small number of tables is involved. An additional tions). A similar encoding happens for the target tableau, except that
constraint (which is true for the actual data) could be used to infer an ∃ clause is used instead of ∀. Finally, the such that clause encodes
that the two occurrences of ExpSet correspond to the same exper- all the correspondences between the source and the target that are
iment set instance. Clio’s chasing engine includes such an inference covered. (Only five of them are shown in the figure.)
mechanism. However, such a constraint is hard to extract in practice A more complex mapping is m3 that is obtained from the pair
(it is not a key, but a complicated dependency). The price to pay for (S4 , T4 ), which is covered by all the correspondences shown in Fig-
not having such a constraint will be further ambiguity that remains to ure 2. An additional complication in generating this mapping arises
be solved during mapping generation (to be described shortly). from the fact that the correspondences that map EXPERIMENTSET
For the target schema, the tableaux are more complex, due to nest- columns to exp set elements/attributes have multiple interpreta-
ing and also to the fact that predicates (such as join conditions) can tions, because each correspondence can match either the first occur-
have a context. While tableaux T1 , T2 and T3 are straight paths to rence or the second occurrence of ExpSet in S4 . This ambiguity
set-type elements (e.g., exp factor), the tableau T4 , intended to is resolved by generating all possible interpretations (e.g., all these
denote the collection of treatment factors treat factor, also con- correspondences match the first occurrence of ExpSet, or all match
tains a join with exp factor. Moreover, the join is relative to a the second occurrence, or some match the first occurrence and some
given instance of exp set. The reason for this is the existence of match the second occurrence). A user would then have to go through
a keyref constraint that associates every treat factor element all these choices and select the desired semantics. The default choice
with an exp factor element, within the same exp set instance. that we provide is the one in which all the ambiguous correspon-
Such keyref constraints, part of the XML Schema specification, can dences match the first choice (e.g., the first occurrence of ExpSet).
be easily specified by putting the constraint on the correct element in For this example, all the different interpretations are in fact equiv-
the hierarchy (exp set instead of the root, for this example). alent, since the two occurrences of ExpSet represent the same in-
To represent tableaux such as T4 and S4 unambiguously, Clio uses stance, due to the constraint discussed earlier.
an internal notation based on: (1) generators, which are used to bind In Clio, the tableaux that are constructed based on chasing form
variables to individual elements in sets, and (2) conditions. Path ex- only the basic (default) way of constructing mappings. Users have
pressions, which can be absolute or relative to the bound variables, the option of creating additional mappings through the mapping edi-
can appear in both the generators and the conditions. As an example, tor. In each schema, a new tableau can be specified by selecting the
the internal form of T4 is shown below. needed collections and then creating predicates on those collections.
Generators: Each such tableau can then participate in the algorithm for mapping
t0 ∈ GEML/exp set, t1 ∈ t0 /exp set header/treatment list/treatment, generation. In other words, each tableau will be paired with all the
t2 ∈ t1 /treat factor list/treat factor, tableaux in the opposite schema to reach all possible mappings, based
t3 ∈ t0 /exp set header/exp factor list/exp factor ; on covered correspondences.
Conditions:
t2 /factor id = t3 /id Mapping language To the language that we have just described, we
must add an additional construct: Skolem functions. These functions
The use of collection-bound variables (rather than that of arbi- can explicitly represent target elements for which no source value is
trarily bound variables) has the advantage that the resulting notation given. For example, the mapping m1 of Figure 4 will not specify
2
also called logical relations in [7]. However, the term tableaux, a value for the @id attribute under exp factor (because there is
used in classical papers on the chase, was the one widely used during no correspondence to map into @id). To create a unique value for
the development of Clio. Hence, we stand by it here. this attribute, which is required by the target schema, a Skolem func-
1. for $x0 in $doc0/GENEX/EXPERIMENTFACTOR,
2. $x1 in $doc0/GENEX/EXPERIMENTSET
as 1–4 except for the variable renaming), and, 2) since it appears
3. where nested within exp set, that such tuples join with the current tuple
4. $x1/ES_PK/text() = $x0/ES_FK/text() from the outer part of the query to create the proper grouping (lines
5. return
6. <exp_set>
19–23 – requiring that the values for exp set in the inner tuple be
7. { attribute id { $x1/ES_PK } } the same as in the outer tuple). The grouping condition in lines 19–23
8. { attribute name { $x1/NAME } } does not appear anywhere in logical mapping m1 . This is computed
9. { attribute local_accession { $x1/LOCAL_ACCESSION } } in the query graph when the shape of the target schema is taken into
10. { attribute release_date { $x1/RELEASE_DATE } }
11. <exp_set_header> consideration. Finally, lines 25–31 produce an actual exp factor
12. <biology_desc> { $x1/BIOLOGY_DESC/text() } </biology_desc> element. Of particular interest, line 30 creates a value for the id el-
13. <analysis_desc> { $x1/ANALYSIS_DESC/text() } </analysis_desc> ement. No correspondence exists for the id element and, thus, there
14. <exp_factors_list> {
15 for $x0L1 IN $doc0/GENEX/EXPERIMENTFACTOR, is no value for it in m1 . However, since id is a required target ele-
16. $x1L1 IN $doc0/GENEX/EXPERIMENTSET ment, the query generator produces a Skolem value for id based on
17. where the dependency information stored in the query graph.
18. $x1L1/ES_PK/text() = $x0L1/ES_FK/text() and
19. $x1/BIOLOGY_DESC/text() = $x1L1/BIOLOGY_DESC/text() and
20. $x1/ANALYSIS_DESC/text() = $x1L1/ANALYSIS_DESC/text() and
21. $x1/NAME/text() = $x1L1/NAME/text() and 3. PRACTICAL CHALLENGES
22. $x1/LOCAL_ACCESSION/text() = $x1L1/LOCAL_ACCESSION/text() and In this section, we present several of the implementation and op-
23. $x1/RELEASE_DATE/text() = $x1L1/RELEASE_DATE/text()
24. return timization techniques that we developed in order to address some of
25. <exp_factor> the many challenging issues in mapping and query generation.
26. { attribute factor_name { $x0L1/FACTOR_NAME } }
27. { attribute factor_units { $x0L1/FACTOR_UNITS } }
28. { attribute major_category { $x0L1/MAJOR_CATEGORY } }
3.1 Scalable Incremental Mapping Generation
29. { attribute minor_category { $x0L1/MINOR_CATEGORY } } One of the features of the basic mapping generation algorithm
30. <id>
{"SK6(",$x1L1/BIOLOGY_DESC/text(),$x1L1/ANALYSIS_DESC/text(),…}</id>
is that it enumerates a priori all possible “skeletons” of mappings,
31. </exp_factor> that is, pairs of source and target tableaux. In a second phase (in-
32. } </exp_factors_list> sertion/deletion of correspondences), mappings are generated, based
33. </exp_set_header> on the precomputed tableaux, as correspondences are added in or
34. </exp_set>
removed. This second phase must be incremental in that, after the
insertion of a correspondence (or of a batch of correspondences) or
Figure 5: XQuery fragment for one of the logical mappings. after the removal of a correspondence, a new mapping state must be
efficiently computed based on the previous mapping state. This is an
tion can be generated (and the siblings and ancestors of @id, such as important requirement since one of the main uses of Clio is interac-
factor name and biology desc, which contain a source value, tive mapping generation and editing. To achieve this, the main data
will appear as arguments of the function). In general, the gener- structure in Clio (the mapping state) is the list of skeletons, where
ation of Skolem functions could be postponed until query genera- a skeleton is a pair of tableaux (source and target) together with all
tion (see Section 2.2) and the schema mapping language itself could the inserted correspondences that match the given pair of tableaux.
avoid Skolem functions. However, to be able to express mappings When correspondences are inserted or deleted, the relevant skeletons
that arise from mapping composition (e.g., a mapping that is equiva- are updated. Furthermore, at any given state, only a subset of all
lent to the sequence of two consecutive schema mappings), functions the skeletons is used to generate mappings. The other skeletons are
must be part of the language [2]. Clio has been recently extended deemed redundant (although they could become non-redundant as
to support mapping composition, an important feature of metadata more correspondences are added4 ). This redundancy check, which
management. Hence, the schema mapping language used in Clio in- we describe next, can significantly reduce the amount of irrelevant
cludes Skolem functions; the resulting language is a nested relational mappings that a user has to go through.
variation of the language of second-order tgds of Fagin et al [2]. Redundancy check A tableau T1 is a sub-tableau of a tableau T2
if there is a one-to-one mapping of the variables of T1 into the vari-
2.2 Query Generation ables of T2 , so that all the generators and all the conditions of T1
Our basic query generation operates as follows. Each logical map- become respective subsets of the generators and conditions of T2 . A
ping is compiled into a query graph that encodes how each target pair (Si , Tj ) of source and target tableaux is a sub-skeleton of a sim-
element/attribute is populated from the source-side data. For each ilar pair (Si′ , Tj′ ) if Si is a sub-tableau of Si′ and Tj is a sub-tableau
logical mapping, query generators walk the relevant part of the target of Tj′ . The sub-tableaux relationships in each of the two schemas
schema and create the necessary join and grouping conditions. The as well as the resulting sub-skeleton relationship are also precom-
query graph also includes information on what source data-values puted in the first phase, to speed up the subsequent processing that
each target element/attribute depends on. This is used to generate occurs in the second phase. When correspondences are added, in
unique values for required target elements as well as for grouping. the second phase, if (Si , Tj ) is a sub-skeleton of (Si′ , Tj′ ) and the
In Figure 5, we show the XQuery fragment that produces the target set C of correspondences covered by (Si′ , Tj′ ) is the same as the set
<exp set> elements as prescribed by logical mapping m1 3 . The of correspondences covered by (Si , Tj ), then the mapping based on
query graph encodes that an <exp set> element will be generated (Si′ , Tj′ ) and C is redundant. Intuitively, we do not want to use large
for every tuple produced by the join of the source tables EXPERIMENT- tableaux unless more correspondences will be covered. For exam-
FACTOR and EXPERIMENTSET (see lines 1–2 in m1 ). Lines 1–4 in ple, suppose that we have inserted only two correspondences, from
Figure 5 implement this join. Lines 7–10 output the attributes within BIOLOGY DESC and ANALYSIS DESC of EXPERIMENTSET to
exp set and implement lines 3–5 of m1 . Then, we need to pro- biology desc and analysis desc under exp set. These
duce the repeatable exp factor elements. The query graph pre- two correspondences match on the skeleton (S2 , T2 ) where S2 and
scribes two things about exp factor: 1) that it will be generated T2 are the tableaux in Figure 3 involving experiment sets and ex-
for every tuple that results from the join of EXPERIMENTFACTOR periment factors. However, this skeleton and the resulting mapping
and EXPERIMENTSET (lines 15–18 in the query – they are the same are redundant, because there is a sub-skeleton (S1 , T1 ) that covers
3 4
The complete XQuery is the concatenation of all the fragments that And vice-versa, non-redundant skeletons can become redundant as
correspond to all the logical mappings. correspondences are removed.
the same correspondences, where S1 and T1 are the tableaux in Fig- els of sets), but we set no limit on the total number of precomputed
ure 3 involving experiment sets. Until a correspondence that maps tableaux, when loading a schema. This experiment gives us a lower
EXPERIMENTFACTOR is added, there is no need to generate a logi- bound estimation on the amount of time and memory that the basic
cal mapping that involves EXPERIMENTFACTOR. algorithm (that precomputes all the tableaux) requires. In the second
The hybrid algorithm The separation of mapping generation into experiment, we control the nesting level of the precomputed tableaux
the two phases (precomputation of tableaux and then insertion of cor- (also, maximum 6) and the total number of precomputed tableaux
respondences) has one major advantage: when correspondences are (maximum 110 per schema). This experiment shows the actual im-
added, no time is spent on finding associations between the elements provement that the hybrid algorithm achieves.
being mapped. All the basic associations are already computed and For the first experiment, we looked at the time to load the MAGE-
encoded in the tableaux; the correspondences are just matched on ML schema on one side only (as source). This includes precomput-
the relevant skeletons, thus speeding up the user addition of corre- ing the tableaux as well as the sub-tableaux relationship. (The time
spondences in the GUI. There is also a downside: when schemas are to compile, in its entirety, the types of MAGE-ML, into the nested re-
large, the number of tableaux can also become large. The number of lational model poses no problem; it is less than 1 second.) We were
skeletons (which is the product of the numbers of source and, respec- able to precompute all (1030) the tableaux that obey the nesting level
tively, target tableaux) is even larger. A significant amount of time limit, in about 2.6 seconds. However, computing the sub-tableaux re-
is then spent in the preprocessing phase to compute the tableaux and lationship (checking all pairs of the 1030 tableaux for the sub-tableau
skeletons as well as the sub-tableaux and sub-skeleton relationships. relationship) takes 74 seconds. The total amount of memory to hold
A large amount of memory may be needed to hold all the data struc- the necessary data structures is 335MB. Finally, loading the MAGE-
tures. Moreover, some of the skeletons may not be needed (or at least ML schema on the target side causes the system to run out of memory
not until some correspondence is added that matches them). (512MB were allocated).
A more scalable solution, which significantly speeds up the initial For the second experiment, we also start by loading the MAGE-
process of loading schemas and precomputing tableaux, without sig- ML schema on the source side. The precomputation of the tableaux
nificantly slowing down the interactive process of adding and remov- (116 now) takes 0.5 seconds. Computing the sub-tableaux relation-
ing correspondences, is the hybrid algorithm. The main idea behind ship (checking all pairs of the 116 tableaux to record if one is a sub-
this algorithm is to precompute only a bounded number of source tableau of another) takes 0.7 seconds. The total amount of mem-
tableaux and target tableaux (up to a certain threshold, such as 100 ory to hold the necessary data structures is 163MB. We were then
tableaux in each schema). We give priority to the top tableaux (i.e., able to load the MAGE-ML schema on the target side in time that is
tableaux that include top-level set-type elements without including similar to that of loading the schema on the source side. The sub-
the more deeply nested set-type element). When a user interacts with sequent computation of the sub-skeleton relationship (checking all
the Clio GUI, she would usually start by browsing the schemas from pairs of the 13456 = 116 x 116 skeletons to record whether one is a
the top and, hence, by adding top-level correspondences, that will sub-skeleton of another) takes 35 seconds. The amount of memory
match the precomputed skeletons. needed to hold everything at this point is 251MB. We then measured
However, correspondences between elements that are deeper in the the performance of adding correspondences. Adding a correspon-
schema trees may fail to match on any of the precomputed skeletons. dence for which there are precomputed matching skeletons is 0.2
We then generate, on the fly, the needed tableaux based on the end- seconds. (This includes the time to match and the time to recompute
points of such a correspondence. Essentially, we generate a source the affected logical mappings.) Removing a correspondence requires
tableau that includes all the set-type elements that are ancestors of the less time. Adding a correspondence for which no precomputed skele-
source element in the correspondence (and similarly for the target). ton matches and for which new tableaux and a new skeleton must be
The tableaux are then closed under the chase, thus including all the computed on the fly takes about 1.3 seconds. (This also includes the
other schema elements that are associated via foreign keys. Next, the time to incrementally recompute the sub-tableaux and sub-skeleton
data structures holding the sub-tableaux and the sub-skeleton rela- relationships.) Overall, we found the performance of the hybrid al-
tionships are updated, incrementally, now that the tableaux and a new gorithm to be quite acceptable and we were able to easily generate
skeleton have been added. The new correspondence will then match 30 of the (many!) logical mappings. Generating the executable script
the newly generated skeleton, and a new mapping can be generated. (query) that implements the logical mappings takes, additionally, a
Overall, the performance of adding a correspondence takes a hit, but few seconds. To give a feel of the complexity of the MAGE-ML
the amount of tableaux computation is limited locally (surrounding transformation, the executable script corresponding to the 30 logical
the end-points) and is usually quite acceptable. As a drawback, the mappings is 25KB (in XQuery) and 131KB (in XSLT).
algorithm may lose its completeness, in that there may be certain The Clio implementation is in Java and the experiments were run
associations between schema elements that will no longer be consid- on a 1600MHz Pentium processor with 1GB main memory.
ered (they would appear if we were to compute all the tableaux as in
the basic algorithm). Still, this is a small price to pay, compared to
the ability to load and map (at least partially) two complex schemas.
3.2 Query Generation: Deep Union
Performance evaluation: mapping MAGE-ML We now give an There are two major drawbacks that the query generation algo-
idea of the effectiveness of the hybrid generation algorithm by illus- rithm described in Section 2.2 suffers from: there is no duplicate
trating it on a mapping scenario that is close to worst case in practice. removal within and among query fragments, and there is no group-
We load the same complex XML schema on both sides and experi- ing of data among query fragments. For instance, suppose we are
ment with the creation of the identity mapping. The schema that we trying to create a list of orders with a list of items nested inside. As-
consider is the MAGE-ML schema 5 , intended to provide a standard sume the input data comes from a simple relational table Orders
for the representation of microarray expression data that would facil- (OrderID,ItemID). If the input data looks like {(o1 ,i1 ), (o1 ,i2 )},
itate the exchange of microarray information between different data our nested query solution produces the following output data: {(o1 ,(i1 ,
systems. MAGE-ML is a complex XML schema: it features many i2 )), (o1 ,(i1 ,i2 ))}. The grouping of items within each order is what
recursive types, contains 422 complex type definitions and 1515 ele- the user expected, but users may reasonably expect that only one in-
ments and attributes, and is 172KB in size. stance of o1 appears in the result. Even if we eliminate duplicates
We perform two experiments. In the first one, we control the from the result of one query fragment, our mapping could result in
nesting level of the precomputed tableaux (maximum 6 nested lev- multiple query fragments, each producing duplicates or extra infor-
mation that needs to be merged with the result of another fragment.
5
http://www.mged.org/Workgroups/MAGE/mage.html For example, assume that a second query fragment produces a tuple
WITH ExpFactorXML AS -- Q4: Add XML tags to the data from Q2.
ExpSetFlat AS -- Q1: The join and union of the relational data for ExpSet (SELECT
(SELECT DISTINCT x0.InSet AS InSet,
x1.BIOLOGY_DESC AS exp_set_exp_set_header_biology_desc, xml2clob(
x1.ANALYSIS_DESC AS exp_set_exp_set_header_analysis_desc, xmlelement(name "exp_factor",
x0.ES_FK AS exp_set_id,
… xmlattribute(name "factor_name", x0.exp_factor_factor_name),
VARCHAR('Sk_GEML_2(' || x1.BIOLOGY_DESC || x1.ANALYSIS_DESC || … || ')') AS ClioSet0, xmlelement(name "id", x0.exp_factor_id),
VARCHAR('Sk_GEML_3(' || x1.BIOLOGY_DESC || x1.ANALYSIS_DESC || … || ')') AS ClioSet1 …
FROM GENEX.EXPERIMENTFACTOR x0, GENEX.EXPERIMENTSET x1 )) AS XML
WHERE x0.ES_FK = x1.ES_PK FROM ExpFactorFlat x0),
UNION
SELECT DISTINCT TreatmentLevelXML AS -- Q5: Add XML tags to the data from Q3.
x1.BIOLOGY_DESC AS exp_set_exp_set_header_biology_desc, (SELECT
x1.ANALYSIS_DESC AS exp_set_exp_set_header_analysis_desc, x0.InSet AS InSet,
x0.ES_FK AS exp_set_id, xml2clob(
… xmlelement(name "treatment",
VARCHAR('Sk_GEML_2(' || x1.BIOLOGY_DESC || x1.ANALYSIS_DESC || … || ')') AS ClioSet0,
xmlattribute(name "id", x0.treatment_id),
VARCHAR('Sk_GEML_3(' || x1.BIOLOGY_DESC || x1.ANALYSIS_DESC || … || ')') AS ClioSet1
FROM GENEX.TREATMENTLEVEL x0, GENEX.EXPERIMENTSET x1
xmlattribute(name "treatment_name", x0.treatment_treatment_name)
WHERE x0.ES_FK = x1.ES_PK), )) AS XML
ExpFactorFlat AS -- Q2: The join of relational data for ExpFactor FROM TreatmentLevelFlat x0),
(SELECT DISTINCT
-- Q6: Combines the results of Q1, Q4, and Q5 into one XML document.
VARCHAR('SK29(' || x0.FACTOR_NAME || … || ')') AS exp_factor_id,
x0.FACTOR_NAME AS exp_factor_factor_name,
SELECT xml2clob(xmlelement (name "exp_set",
… xmlattribute (name "id", x0.exp_set_id),
VARCHAR('Sk_GEML_2(' || x1.BIOLOGY_DESC || x1.ANALYSIS_DESC || … || ')') AS InSet …
FROM GENEX.EXPERIMENTFACTOR x0, GENEX.EXPERIMENTSET x1 xmlelement (name "exp_set_header",
WHERE x0.ES_FK = x1.ES_PK), xmlelement (name "biology_desc", x0.exp_set_exp_set_header_biology_desc),
TreatmentLevelFlat AS -- Q3: The join of relational data for TreatmentLevel xmlelement (name "analysis_desc", x0.exp_set_exp_set_header_analysis_desc),
(SELECT DISTINCT xmlelement (name "exp_factors_list",
VARCHAR('SK109(' || x0.NAME || ',' … || ')') AS treatment_id,
(SELECT xmlagg (x1.XML)
x0.NAME AS treatment_treatment_name,
FROM ExpFactorXML x1
VARCHAR('Sk_GEML_4 (' || 'SK110(' || x0.NAME || x1.RELEASE_DATE || … || ')') AS ClioSet0,
VARCHAR('Sk_GEML_3(' || x1.BIOLOGY_DESC || x1.ANALYSIS_DESC || … || ')') AS InSet WHERE x1.InSet = x0.ClioSet0)),
FROM GENEX.TREATMENTLEVEL x0, GENEX.EXPERIMENTSET x1 xmlelement (name "treatment_list",
WHERE x0.ES_FK = x1.ES_PK), (SELECT xmlagg (x1.XML)
FROM TreatmentLevelXML x1
WHERE x1.InSet = x0.ClioSet1)))
Figure 6: SQL/XML script, relational part. )) AS XML
FROM ExpSetFlat x0
{(o1 ,(i3 ))}. We would expect this tuple to be merged with the previ-
Figure 7: SQL/XML script continued, XML construction part.
ous result and produce only one tuple for o1 with three items nested
inside. We call this special union operation deep-union. to remove duplicate values.
We illustrate the algorithm by showing the generated query in the In the second half of the script, the query fragments Q4 and Q5
case of SQL/XML, a recent industrial standard that extends SQL with perform the appropriate XML tagging of the results of Q2 and Q3.
XML construction capabilities. For the example, we assume that we Finally, Q6 tags the exp set result of Q1 and, additionally, joins
are only interested in generating the transformation for two of our with Q4 and Q5 using the created set-IDs, in order to nest all the
logical mappings: m1 (mapping experiment sets with their associ- corresponding exp factor and treatment elements.
ated experiment factors) and m2 (which is similar to m1 and maps
experiment sets with their associated treatment levels).
The generated SQL/XML script can be separated in two parts. The
4. REMAINING CHALLENGES
first part (shown in Figure 6) generates a flat representation of the out- There remain a number of open issues regarding scalability and
put in which a collection of tuples is represented by a system gener- expressiveness of mappings. Complex mappings sometimes need a
ated ID, and each tuple contains the ID of the collection it is supposed more expressive correspondence selection mechanism than that sup-
to belong to. The purpose of the second part (Figure 7) is to recon- ported by Clio. For instance, deciding which group of correspon-
struct the hierarchical structure of the target by joining tuples based dences to use in a logical mapping may be based on a set of pred-
on their IDs (i.e., joining parent collections with the corresponding icates. We are also exploring the need for logical mappings that
children elements based on IDs). The final result is free of duplicates nest other logical mappings inside. Finally, we are studying mapping
and merged according to the deep union semantics. adaptation issues that arise when source and target schemas change.
Briefly, Q1 joins EXPERIMENTFACTOR with EXPERIMENTSET
and will be used to populate the atomic components at the exp set 5. REFERENCES
level in the target (thus, not including the atomic data that goes under [1] P. Bernstein. Applying Model Management to Classical Meta Data
Problems. In CIDR, 2003.
exp factor and treatment). Two set-IDs are generated in Q1
[2] R. Fagin, P. Kolaitis, L. Popa, and W.-C. Tan. Composing Schema
(under the columns ClioSet0 and ClioSet1), for each different Mappings: Second-Order Dependencies to the Rescue. In PODS, 2004.
set of values that populate the atomic components at the exp set [3] R. Fagin, P. G. Kolaitis, R. J. Miller, and L. Popa. Data Exchange:
level. The first one, ClioSet0, will be used to group exp factor Semantics and Query Answering. In ICDT, 2003.
elements under exp set, while ClioSet1 will be used to group [4] M. Lenzerini. Data Integration: A Theoretical Perspective. In PODS,
treatment elements. The values for the set-IDs are generated as 2002.
strings, by using two distinct Skolem functions that depend on all the [5] S. Melnik, P. A. Bernstein, A. Halevy, and E. Rahm. Supporting
atomic data at the exp set level. The atomic data for exp factor Executable Mappings in Model Management. In SIGMOD, 2005.
and treatment are created by Q2 and Q3, respectively. In both [6] R. J. Miller, L. M. Haas, and M. A. Hernández. Schema Mapping as
Query Discovery. In VLDB, 2000.
cases, a set-ID (named InSet) is created to capture what experi-
[7] L. Popa, Y. Velegrakis, R. J. Miller, M. A. Hernández, and R. Fagin.
ment set the data belongs to. The main idea here is that, as long as Translating Web Data. In VLDB, 2002.
the values that go into the atomic components at the exp set level [8] E. Rahm and P. A. Bernstein. A Survey of Approaches to Automatic
are the same, the InSet set-ID will match the set-ID stored under Schema Matching. The VLDB Journal, 10(4):334–350, 2001.
ClioSet0 (in the case of Q2) or ClioSet1 (in the case of Q3). [9] N. C. Shu, B. C. Housel, R. W. Taylor, S. P. Ghosh, and V. Y. Lum.
On a different note, we remark that all queries that appear in the first EXPRESS: A Data EXtraction, Processing, and REStructuring System.
half of the script (e.g., Q1, Q2, and Q3) use the DISTINCT clause TODS, 2(2):134–174, 1977.