程序代写代做 cache graph algorithm IOS ER C database Functional Dependencies Contents:

Contents:
CS W186 Spring 2019 Midterm 2
Do not turn this page until instructed to start the exam.
• You should receive one double-sided answer sheet and a 20-page exam packet. • The midterm has 6 questions, each with multiple parts.
• The midterm is worth a total of 76 points.
Taking the exam:
• You have 110 minutes to complete the midterm.
• All answers should be written on the answer sheet. The exam packet will be collected but not graded.
• For each question, place only your final answer on the answer sheet; do not show work.
• For multiple choice questions, please fill in the bubble or box completely as shown on the left below. Do not mark the box with an X or checkmark.
• Use the blank spaces in your exam for scratch paper.
Aids:
• You are allowed two handwritten 8.5” × 11” double-sided pages of notes.
• The only electronic devices allowed are basic scientific calculators with simple numeric readout. No
graphing calculators, tablets, cellphones, smartwatches, laptops, etc.
Grading Notes:
• All IOs must be written as integers. There is no such thing as 1.04 IOs – that is actually 2 IOs.
• 1 KB = 1024 bytes. We will be using powers of 2, not powers of 10
• Unsimplified answers, like those left in log format where simplification to integers is possible, will receive a point penalty.

1 Joins (18 points)
Consider the following relations
CREATE TABLE Customers (
cid INTEGER PRIMARY KEY,
name CHAR(30),
address CHAR(30)
);
CREATE TABLE Orders (
oid INTEGER PRIMARY KEY,
cid INTEGER REFERENCES Customers(cid),
item_name CHAR(20),
item_count INTEGER
);
In this problem, we will consider executing the following query using various join algorithms.
SELECT *
FROM Customers C, Orders O
WHERE C.cid = O.cid;
Assume that
• Orders has [O] = 100 pages, with pO = 50 tuples per page.
• Customers has [C] = 40 pages, with pC = 25 tuples per page.
• We have an Alternative 2 B+ tree index on Orders.cid with height h = 2 (recall that a tree with only the root has height 0).
• Each customer has at least one order, and Orders.cid is uniformly distributed.
• Our buffer has size B = 10, unless noted otherwise.
• For index nested loops join, assume the cost model from lecture, where we do not cache index pages.
• Unless otherwise noted, each of the following parts are to be answered independently, i.e. assume the query is executed from scratch for each part.
• The files are not already sorted.
1. (2 points) In the best case, what is the I/O cost of a block nested loops join?
2. (2 points) Assume the index on Orders.cid is clustered. In the best case, what is the I/O cost of an index nested loops join?
Solution: [C]+⌈ [C] ⌉·[O]=40+5·100=540IOs B−2
Solution: From the assumptions, we have that each customer has 5 matches in Orders. In the best case, all 5 matches appear on the same page because the index is clustered. The cost is
[C] + |C| · (3 + 1) = 40 + 4000 = 4040 I/Os
Page 2 of 20

3. (2 points) Assume the index on Orders.cid is unclustered. In the worst case, what is the I/O cost of an index nested loops join?
4. (2 points) In the best case, what is the I/O cost of a (unoptimized) sort-merge join?
Solution: In the worst case, all 5 matches are on separate pages for each tuple in Customers. So the I/O cost is
[C] + |C| · (3 + 5) = 40 + 8000 = 8040 IOs
Solution: The cost is
􏰈 [C] 􏰉 􏰈 [O] 􏰉
2· 1+⌈logB−1⌈ B ⌉⌉ ·[C]+2· 1+⌈logB−1⌈ B ⌉⌉ ·[O]+[O]+[C]=160+600+100+40 = 900 I/Os
5. (3 points) What is the minimum buffer size B in order to perform the optimized sort-merge join? Recall that the optimized sort-merge join uses 2 passes: 1 sorting pass, and 1 pass for both merging and joining.
Solution: In order to do the sort-merge join refinement, we need B > ⌈[O]⌉ + ⌈[C]⌉
BB The minimum B that satisfies this inequality is B = 13.
Partial credit was given for the approximation
B ≥ 􏰇[O] + 􏰇[C]
which yields a value of B = 17.
6. (2 points) Assuming that our buffer is big enough, what is the I/O cost of the optimized sort-merge join described in the previous part?
7. (1 point) Assume we do a grace hash join and that Customers is the building relation, i.e. we build the in-memory hash table using Customers. Assume after the first partitioning pass, we partition Customers into the following partitions (assume all pages are full):
• PC1 : 10 pages • PC2 : 9 pages
Solution: 3 · ([O] + [C]) = 420 I/Os
Page 3 of 20

• PC3 : 9 pages • PC4 : 8 pages • PC5 : 4 pages
with the remaining partitions having 0 pages. What is the size (in pages) after the first partitioning pass of the partitions PO1, PO2, PO3, PO4, PO5 respectively (as a tuple of five integers)? Here POi denotes the i-th partition of Orders after the first partitioning pass, assuming the same ordering of partitions as the PCi partitions described above.
A. (50, 45, 45, 40, 20) B. (25, 22, 22, 20, 10) C. (12, 12, 12, 12, 12)
D. (25, 23, 23, 20, 10)
E. (20, 20, 20, 20, 20)
Solution: D
Since each customer has 5 matches in Orders, and there are 50 tuples per page of Orders vs. 25 tuples per page of Customers, we have that (remember the number of pages must be an integer):
• PO1 :10·25·5=1250tuples=25pages • PO2 :9·25·5=1125tuples=23pages • PO3 :9·25·5=1125tuples=23pages • PO4 :8·25·5=1000tuples=20pages • PO5 :4·25·5=500tuples=10pages
Therefore the answer is (25, 23, 23, 20, 10).
8. (4 points) Assume that the hash function used for recursive partitioning hr partitions each partition of Customers uniformly (or as close to uniformly as possible) among 2 new partitions. What is the I/O cost of grace hash join from the previous part? Recall that we only perform recursive partitioning when necessary, i.e. we further divide only the partitions that are too large for the build and probe phase.
Solution: After the first pass, we must do recursive partitioning for all Customers partitions of size greater than B − 2, and the corresponding Orders partitions. This means we must recursively partition PCi, POi for i ∈ {1, 2, 3}. After recursive partitioning, since hr divides the data uniformly among 2 new partitions, all partitions of Customers are of size less than or equal to B − 2.
The only place where we have to be careful are the partitions of Orders with non-full pages, PO2, PO3. We can quickly verify that these partitions will be partitioned by hr into two partitions of 12 pages each by doing
⌈(⌊(9 · 25)/2⌋ · 5)/50⌉ = 12 ⌈(⌈(9 · 25)/2⌉ · 5)/50⌉ = 12
Page 4 of 20

Thus, the total cost is:
[O]+2·(8+4)+2·(10+9+9)+2·(10+10+10)+[C]+2·(20+10)+2·(25+23+23)+2·(26+24+24) = 630 I/Os.
Partial credit was given for
2 · ([O] + [C]) + 2 · (10 + 9 + 9 + 25 + 23 + 23) + [O] + [C] = 618 I/Os.
No partial credit was given if the previous part was calculated incorrectly.
Page 5 of 20

2 Parallelism (12 points)
For this question, assume the following:
• Assume that we have three tables: Teachers(name, salary, course), Students(name, SID, course,
grade), Employees(name, eid, company, salary)
• [T] = 10 pages
• [S] = 10000 pages
• [E] = 100000 pages
• 1page=4KB
• We have 4 machines with 52 buffer pages each
• Assume each I/O takes 1 ms. There are 1000ms in a second.
• All questions are to be considered independently of each other.
1. (2 points) Assume Students is round-robin partitioned across the 4 machines, and that Teachers lies entirely on Machine 1. What is the minimum network cost (in KB) incurred from performing a join between Teachers and Students?
2. (2 points) If Students is round-robin partitioned across the 4 machines and Employees is range-partitioned on the name key across 4 machines, what would be the worst-case network cost be (in KB) to perform
a sort-merge join between Students and Employees on the name key?
3. (2 points) Let us assume Students and Employees are both hash-partitioned on the name key using the same hash function across the 4 machines and that, due to key skew, machine 1 has 50% of Students on it, with the remaining 50% of the Students table distributed evenly among machines 2, 3, and 4. Assume Employees is distributed uniformly across all 4 machines. If we were to execute a block nested loop join between Students and Employees on each machine, how much time would be taken, in seconds, to complete this? Ignore the time it takes to write the output.
Solution: Broadcast join: 10*3*4 KB = 120 KB. Because Teachers lies entirely on machine 1, we must send it to the other 3 machines and then execute the join.
Solution: 10000 * 4KB = 40000 KB; for each machine, process all records and send it to the appropriate machine for the appropriate partitioning. In the absolute worst case, no Students record would be on the correct machine and they would all need to be sent over the network to the right machine.
Solution: Machine 1 gets 5000 pages from Students and 25000 pages from Employees; this machine will take the longest, as it is the bottleneck.
(5000 + ⌈ 5000 ⌉ ∗ 25000) ∗ 1 ms ∗ 1 s = 2505 seconds 50 1000 ms
Page 6 of 20

4. (2 points) Assuming Students is range-partitioned across the 4 machines with 40% of Students going to machine 1 and the remaining 60% split evenly across machines 2, 3, and 4. How long, in seconds, would it take to parallel sort the Students table?
5. (2 points) Now, assume Students is range-partitioned on the name key evenly across the 4 machines. How long, in seconds, would it take to parallel sort the Students table on this key?
6. (2 points) Now, we wish to calculate the square root of the sum of the squared salary of each row in Employees, i.e. SELECT SQRT(SUM(salary * salary)) FROM Employees. Mark a global and lo- cal aggregate function that would allow us to efficiently calculate SQRT(SUM(salary * salary)) via hierarchical aggregation.
Solution: Machine 1 gets 4000 pages, so this is the bottleneck in terms of speed.
After 1 merge pass, we have ⌈4000⌉ ≈ 80 runs, so we need 3 passes total (2 merge passes). Time
51
taken: 2∗k∗4000∗.001 s = 24000 I/Os * .001 s = 24 seconds
Solution: Each machine gets 2500 pages.
After 1 merge pass, we have ⌈ 2500 ⌉ < 50 runs, so we need 2 passes total (1 merge pass). Time taken: 51 2∗k∗2500∗.001 s = 10000 I/Os * .001 s = 10 seconds Local Global A. count(x) B. x C. 􏰆x D. 􏰆x2 E. 􏰇􏰆x F. 􏰇􏰆x2 A. count(x) B. x C. 􏰆x D. 􏰆x2 E. 􏰇􏰆x F. 􏰇􏰆x2 Page 7 of 20 3 Query Optimization (15 points) For questions 1-4, we consider the following schema: CREATE TABLE Guitars( gid INTEGER PRIMARY KEY, brand VARCHAR(50), price INTEGER ); CREATE TABLE Players( pid INTEGER PRIMARY KEY, name VARCHAR(50), age INTEGER ); CREATE TABLE LastPlayed( gid INTEGER REFERENCES Guitars(gid), pid INTEGER REFERENCES Players(pid), date DATE, PRIMARY KEY (gid, pid) ); We make the following assumptions about the distribution of data: • 10 ≤ Players.age < 85 • 1,000 ≤ Guitars.price < 5,000 • Players.pid has 1,000 distinct values • Guitars.gid has 1,000 distinct values • Guitars.brand has 10 distinct values • Players.age is independent from Guitars.brand Consider the following query: SELECT P.name FROM Guitars G, Players P, LastPlayed L WHERE G.gid = L.gid AND P.pid = L.pid AND (P.age < 25 OR G.brand = ’CS186’) AND G.price >= 3000;
Compute the selectivity for each of the following predicates from the WHERE clause. Write only your final answer, as a simple fraction.
1. (1 point) G.gid = L.gid
Solution: 1/1000 Explanation:
Selectivity = 1 / MAX(NKeys(G.gid), NKeys(L.gid)) NKeys(G.gid) = 1000
NKeys(L.gid) ≤ 1000 because of the foreign key referencing G.gid 1/max(1000, x) = 1000, since x ≤ 1000.
Page 8 of 20

2. (1 point) P.pid = L.pid
Solution: 1/1000 Explanation:
• Selectivity = 1 max(NKeys(P.pid), NKeys(L.pid))
• NKeys(P.pid) = 1000
• NKeys(L.pid) <= 1000 because of foreign key referencing P.pid So 1/Max(1000, x) where x <= 1000 = 1/1000. 3. (1 point) G.price >= 3000
Solution: 1/2
Explanation: Selectivity = (High(G.price)−3000) + 1 = 2399/4000
Selectivity = 4999−3000 + 1 = 2399/4000 4999−1000+1 10
(High(G.price)−Low(G.price) + 1) number distinct values of G.price
4. (1 point) P.age < 25 OR G.brand = ’CS186’ Solution: 14/50 = 7/25 Explanation: Selectivity= 1/5 + 1/10 - 1/5*1/10 = 14/50 For the next question, consider the following schema: CREATE TABLE Product( pid INTEGER PRIMARY KEY, name TEXT, price INTEGER Page 9 of 20 ); CREATE TABLE Company( cid INTEGER, pid INTEGER REFERENCES Product(pid), name TEXT, PRIMARY KEY (cid, pid) ); • Product contains 20,000 tuples, and each record is 20 bytes long. • Company contains 1,000 tuples, and each record is 25 bytes long. • Each page can hold 5,000 bytes. • The buffer pool is 102 pages large. • The fill factor for all hash tables is 0.8. • There are no indices. 5. (2 points) Consider all the join algorithms covered so far in this class. What is the minimum estimated I/O cost to execute the following query? Exclude the final write from your solution. SELECT P.name, C.name FROM Product P, Company C WHERE P.pid = C.pid Solution: 85 Explanation: [Product] = 20,000·20 = 80 5,000 [Company] = 1,000·25 = 5 5,000 We can perform a naive in-memory hash join, since [Company] = 5 < fill factor · (B − 2) = 0.8 · (102 − 2) = 80. The I/O cost is [Product] + [Company] = 85. BNLJ also works and yields the same I/O cost. For the next two questions, we consider the following query on the relations R(a,b,c,d,e), S(a,b,c,d,f), T(a,b,c,d), U(a,b,c,d): SELECT * FROM R INNER JOIN S ON R.a = S.a AND R.b = S.b INNER JOIN T ON S.c = T.c AND S.d = T.d INNER JOIN U ON U.c = R.c AND R.a = U.a WHERE R.e > 1000 AND S.f > 0
GROUP BY S.c
ORDER BY R.b LIMIT 10;
6. (2 points) For each of the following joins, mark True if the Selinger optimizer considers it at some point, and False if not.
A. (R ◃▹ S) ◃▹ T
Page 10 of 20

B. (R◃▹U)◃▹(S◃▹T)
C. ((S◃▹U)◃▹R)◃▹T D. ((R◃▹U)◃▹S)◃▹T
7. (4.5 points) Consider the following table while running a pass of the Selinger optimizer’s dynamic pro- gramming algorithm:
Solution: (a) and (d) are considered.
(b) is not a left-deep tree.
(c) contains a cartesian product (S ◃▹ U), which the Selinger optimizer avoids.
Left Relations
Right Relation
Sorted on
I/O cost
Output Size
(a)
{S, T, U}
R
N/A
10,000
4,000
(b)
R.a
12,000
(c)
(R.b, R.a)
11,000
(d)
{R, T, S}
U
N/A
2,000
4,000
(e)
(R.a, R.b)
52,000
(f)
{R, U, T}
S
N/A
22,000
4,000
(g)
S.c
100,000
(h)
{R, U, S}
T
N/A
150,000
4,000
(i)
S.c
110,000
For each row, mark whether the row is retained at the end of the pass.
Solution: (c), (d), (g) are retained.
There are no downstream joins (since this is the final pass), so the only interesting orders arise from
GROUP BY and ORDER BY clauses, which are on S.c and R.b respectively.
We are comparing all the rows with each other, since they all return the set R, S, T, U.
(d) is the minimal cost join and is retained.
(c) is the only join that sorts on R.b, and is retained.
(g) and (i) are the only joins that sort on S.c, and (g) is lower cost, so only (g) is retained.
For each of the following questions, mark either True or False.
8. (0.5 points) The Selinger algorithm will produce an optimal query plan, given a perfect cost estimator.
Solution: False. The Selinger algorithm produces the best query plan out of the space of query plans we are considering given our cost estimation. However, the best query plan might not be left-deep. The Selinger algorithm doesn’t consider cartesian products either, which also limits our plan space.
Page 11 of 20

9. (0.5 points) An output is sorted on column T.c, where T.c is used in a join in the query. Then, we have an interesting order on T.c.
10. (0.5 points) If we only consider BNLJ and do not consider interesting orders while joining n tables, the number of left-deep query plans is O(n!)
11. (0.5 points) True or False: the number of joins considered over the course of the dynamic programming algorithm is always at least exponential in the number of tables we are selecting from.
12. (0.5 points) Consider a relation R which is sorted on some column c and R.c is used in an ORDER BY clause. A page nested loop join with R as the outer relation produces an interesting order of R.c.
Solution: False. It has to be used in a downstream join. If our join condition was T.c = T2.b and we did a sort-merge join, then we would have the output sorted on T.c
Solution: False. The number of query plans is at most cnn!, where c is the maximum number of single-table access operators for a particular relation.
Solution: False, if we have n joins and n cartesian products, we first consider O(n2n ) joins, find
Page 12 of 20
2
22nn the best one, then find the best way to do the 2 cartesian products, which is also O(n2 2 ).
Solution: False, PNLJ does not produce an interesting order.

4 Text Search (8 points)
1. (2 points) For each assertion, fill in the corresponding bubble True or False.
A. In the “Bag of Words” model, we might choose to ignore the word “a” in phrase “a table” because it doesn’t contain much information. This is an example of a stop word.
B. In vector space model, the dimension of vectors depends on the length of the longest document.
C. In text search engine, querying is efficient but updates are not.
D. The IDF part of the TF-IDF value is responsible for favoring repeated terms.
2. (2 points) We have 64 documents, and the following table summarizes the number of documents con- taining a term and the number of occurrences of a term in doc1 for several terms:
What is the TF-IDF for term = “Database” for doc1?. Assume base-2 for any logarithms.
3. (2 points) We are given documents d1 and d2 with a ”bag of words” containing the vocabulary (Dog, Kangaroo, Cat, String, Scratch, Wood). We decide to model them with the following vectors: d1 = [1, 3, 2, 1, 1, 0]⊤ and d2 = [1, 0, 0, 0, 2, 2]⊤. Find the cosine similarity between d1 and d2. Write your answer as a decimal, rounded to two decimal places.
4. (1 point) Suppose that there exist 186 documents about database systems in our corpus. We query for documents about database systems, and our query returns 25 results, 13 of which are actually about database systems. What is the precision? Express your answer as a fraction, you do not need to simplify.
Solution: A, C
B is wrong. The dimension depends on the number of possible terms. D is wrong. Should be unusual terms.
term
# docs containing term
# terms in doc1
Algorithm
4
20
Berkeley
16
17
Computer
1
12
Database
16
21
Solution: 42 Explanation:
21 · log2(64/16) = 42
Solution: cosθ = d1·d2
√ ∥d1∥∥d2∥ √
∥d1∥= 1+9+4+1+1=4,and∥d2∥= 1+4+4=3
= 1·1+2·1 = 0.25 4·3
Page 13 of 20

Solution: precision = true positives / selected items = 13/25
5. (1 point) What is the recall from the previous part (expressed as a fraction)?
Solution: recall = true positives / relevant items = 13 / 186
Page 14 of 20

5 ER Diagrams (11 points)
Welcome to Hogwarts! As a new student you are trying to understand how the school works. You are given the diagram below, but some information is missing! Fill in the diagram using the details below:
• A subject must have at least one student studying it
• A professor is uniquely identified by their name and the subject they teach • Every subject must have at least one professor teaching it
• Each house has exactly one professor who is the head of house.
• Each professor can be the head of at most one house
• Every house must have at least one member
• Every student must be in exactly one house
1. (1 point) ship set?
A. B. C.
D.
2. (1 point) set?
A.
B.
C.
What type of edge should be drawn between the Students entity and the In House relation-
Thin Line Bold Line Thin Arrow Bold Arrow
What type of edge should be drawn between the House entity and the In House relationship
Thin Line
Bold Line
Thin Arrow
Solution: Each student must be in exactly one House. This means that a student must have a key constraint and participation constraint on their relationship with In House
Page 15 of 20

D.
3. (1 point) ship set?
A. B. C.
D.
4. (1 point) set?
A.
B.
C. D.
5. (1 point) A. B. C. D.
6. (1 point) ship set?
A. B.
Bold Arrow
What type of edge should be drawn between the Professor entity and the Teaches relation-
Thin Line Bold Line Thin Arrow Bold Arrow
What type of edge should be drawn between the Subject entity and the Study relationship
Thin Line
Bold Line
Thin Arrow Bold Arrow
What type of relation is Professor to Head of? many-to-many
many-to-one
one-to-many
one-to-one
What type of edge should be drawn between the Professor entity and the Head of relation-
Thin Line Bold Line
Solution: Each House must have at least one student to be considered a House. This means that it requiers a participation constraint. However, a House may have many Students in it so there is no key constraint.
Solution: Professors are uniquely identified by the subject that they teach. This means that they are a weak entity unique identified by the combination of their name and the subject name. Weak entities must have a bold arrow relationship with the strong entity that defines them.
Solution: Each Subject must have at least one student that studies it to be a Subject. This means that it requires a participation constraint. However, a Subject may have many Students studying it so there is no key constraint.
Solution: Each House must have exactly one Professor who is the Head of House. And each Professor may be the Head of House for at most one House. This is a one-to-one relationship.
Page 16 of 20

C. Thin Arrow
D. Bold Arrow
7. (1 point) True or False: Students are required to study one or more Subjects. A. True
B. False
8. (1 point) True or False: Multiple Professors may teach one Subject A. True
B. False
9. (1 point) True or False: The Professor entity is a weak entity A. True
B. False
10. (2 points) In their fifth year, students may be nominated as prefect. Each house may have one female and one male prefect. Which of the following would be the effective way to represent this in the diagram?
A. A new relationship set between Students and House called Prefect with a bold line connecting Student to Prefect and a thin line connecting House to Prefect because it is a many-to-one relationship
B. A new relationship set between Students and House called Prefect with a bold arrow from Student to Prefect and 2 bold arrows from House to Prefect
C. 2 new relationships between Students and House called Female Prefect and Male Prefect with a thin arrow from Student to Female Prefect and a bold arrow from House to Female Prefect. Also a thin arrow from Student to Male Prefect and a bold arrow from House to Male Prefect.
D. This is impossible. You cannot have this type of relationship in an ER diagram
Solution: Each Professor may be the Head of House for at most one House.
Solution: Subjects must be taken by at least one Student, but Students do not have any requirements on the number of courses they must take.
Solution: A Professor must teach exactly one Subject but a Subject may be taught by one or more Professors
Solution: Professors are uniquely identified by the Subject they teach rather than a unique independent identifier such as a name or id. This makes them a weak entity.
Page 17 of 20

Solution: Option A is incorrect because it is not a many-to-one relationship. Student to Prefect is a 2-to-2 relationship Option B is incorrect because it does not guarantee exactly one female and one male Prefect for each house Option C is correct because you can encapsulate this with 2 one-to-one relationships. One for Student to Female Prefect and one from Student to Male Prefect Option D is incorrect but you can represent this in the ER diagram.
Page 18 of 20

6 Functional Dependencies (12 points)
1. (5 points) Decompose R = ABCDEF into BCNF in the order of the following functional dependencies: AE → F, A → B, BC → D, CD → A, CE → D.
Which of the following tables are included in the final decomposition? A. ABC
B. ACD C. AEF D. BCD E. CDE
Solution: AE → F: ABCDEF decomposes into ABCDE and AEF A → B: ABCDE decomposes into ACDE and AB
BC → D: Skip
CD → A: ACDE decomposes into ACD and CDE
CE → D: Skip
Final decomposition: AB, ACD, AEF, CDE
For the next 4 questions, consider relation R = ABCDEF and functional dependencies F = {AB → CF, CE → B, F → D}.
2. (2.5 points) Which of the following are superkeys of this relation? A. AB
B. ABE C. ACE
D. ABCD
E. ABCDE
Solution: AB → ABCDF ABE → ABCDEF
ACE → ABCDEF
ABCD → ABCDF ABCDE → ABCDEF
3. (2.5 points) Which of the following are candidate keys of this relation? A. AB
B. ABE C. ACE
D. ABCD E. ABCDE
Page 19 of 20

Solution: Continuing from the question above, any subset of ABE or ACE does not determine all columns of R, so ABE and ACE are candidate keys. Then ABCDE is not a candidate key.
4. (1 point) True or False: The decomposition of R into ABCE and ABDEF is a dependency preserving decomposition.
5. (1 point) True or False: The decomposition of R into ABCE and ABDEF is a lossless decomposition.
Solution: True. AB → C and CE → B are preserved in ABCE, AB → F and F → D are preserved in ABDEF.
Solution: True. ABCE ∩ ABDEF = ABE and we know that ABE → ABCE from part 2.
Page 20 of 20