Integrating Column-Oriented Storage and Query Processing Techniques Into Management Systems

Pranjal Gupta, Amine Mhedhbi, Semih Salihoglu University of Waterloo {pranjal.gupta, amine.mhedhbi, semih.salihoglu}@uwaterloo.ca

ABSTRACT such, these columnar techniques are relevant for improving We revisit column-oriented storage and query processing the performance and scalability of GDBMSs. techniques in the context of contemporary graph database In this paper, we revisit columnar storage and query pro- management systems (GDBMSs). Similar to column-orien- cessing techniques in the context of GDBMSs. Specifically, ted RDBMSs, GDBMSs support read-heavy analytical work- we focus on an in-memory GDBMS setting and discuss the loads that however have fundamentally different data access applicability of columnar storage and compression techniques patterns than traditional analytical workloads. We first de- for storing different components of graphs [17, 19, 58, 66], rive a set of desiderata for optimizing storage and query and block-based query processing [20, 27]. Despite their sim- processors of GDBMS based on their access patterns. We ilarities, workloads in GDBMS and column-oriented RDBMSs then present the design of columnar storage, compression, also have fundamentally different access patterns. For ex- and query processing techniques based on these desiderata. ample, workloads in GDBMSs contain large many-to-many In addition to showing direct integration of existing tech- joins, which are not frequent in column-oriented RDBMSs. niques from columnar RDBMSs, we also propose novel ones This calls for redesigning columnar techniques in the context that are optimized for GDBMSs. These include a novel list- of GDBMSs. The contributions of this paper are as follows. based query processor, which avoids expensive data copies Guidelines and Desiderata: We begin in Section3 by ana- of traditional block-based processors under many-to-many lyzing the properties of data access patterns in GDBMSs. joins and avoids materializing adjacency lists in intermedi- For example, we observe that different components of data ate tuples, a new data structure we call single-indexed edge stored in GDBMSs can have some structure and the order property pages and an accompanying edge ID scheme, and a in which operators access vertex and edge properties often new application of Jacobson’s bit vector index for compress- follow the order of edges in adjacency lists. This analysis ing NULL values and empty lists. We integrated our tech- instructs a set of guidelines and desiderata for designing the niques into the GraphflowDB in-memory GDBMS. Through physical data layout and query processor of a GDBMS. extensive experiments, we demonstrate the scalability and Columnar Storage: Section4 explores the application of query performance benefits of our techniques. columnar data structures for storing different components of GDBMSs. While existing columnar structures can directly be used for storing vertex properties and many-to-many (n- 1. INTRODUCTION n) edges, we observe that using a straightforward columnar Contemporary GDBMSs are data management software structure, which we call edge columns, to store properties such as Neo4j [9], Neptune [1], TigerGraph [13], and Graph- of n-n edges is suboptimal as it does not guarantee sequen- flowDB [44, 50] that adopt the property graph data model [10]. tial access when reading edge properties in either forward or In this model, application data is represented as a set of ver- backward directions. An alternative, which we call double- tices, which represent the entities in the application, directed indexed property CSRs, can achieve sequential access in both arXiv:2103.02284v1 [cs.DB] 3 Mar 2021 edges, which represent the relationships between entities, directions but requires duplicating edge properties, which and key-value properties on the vertices and edges. can be undesirable as graph-structured data often contain GDBMSs have lately gained popularity to support a wide orders of magnitude more edges than vertices. We then de- range of analytical applications, from fraud detection and scribe an alternative design point, single-directional property risk assessment in financial services to recommendations in pages, that avoids duplication and achieves good locality e-commerce and social networks [57]. These applications when reading properties of edges in one direction and still have workloads that search for patterns in a graph-structured guarantees random access in the other when using a new database, which often requires reading large amounts of edge ID scheme that we describe. Our new ID schemes allow data. In the context of RDBMSs, column-oriented sys- for extensive compression when storing them in adjacency tems [11, 41, 58, 63] employ a set of read-optimized storage, lists without decompression overheads. Lastly, as a new ap- indexing, and query processing techniques to support tradi- plication of vertex columns, we show that single cardinality tional analytical applications, such as business intelligence edges and edge properties, i.e. those with one-to-one (1-1), and reporting, that also process large amounts of data. As one-to-many (1-n) or many-to-one (n-1) cardinalities, are stored more efficiently with vertex columns instead of the structures we described above for n-n edges.

1 name: “peter” name: “alice” age: 17 Columnar Compression: In Section5, we review existing since: 2015 age: 45 Vertex labels - gender: M gender: F columnar compression techniques, such as dictionary encod- PERSON ing, that satisfy our desiderata and can be directly applied doj: 2014 doj: 2019

ORG

to GDBMSs. We next show that existing techniques for Name: “UofT” compressing NULL values in columns from references [17, name: “UW” doj: 2006 estd: 1885

estd: 1934 19] by Abadi et al. lead to very slow accesses to arbitrary since: 2011 Edge labels - doj: 2006 non-NULL values. We then review Jacobson’s bit vector in- since: 1999 FOLLOWS

since: 1992 since: 2006 since: doj: 1980 dex [42, 43] to support constant time rank queries, which has since: since:2009 2012 STUDYAT found several prior applications e.g., recently in a range filter name: “jenny” name: “bob” WORKAT age: 23 age: 54 gender: M structure in databases [61], in information retrieval [37, 52] gender: F since: 2003 ★ doj: date of joining and computational geometry [30, 53]. We show how to en- hance one of Abadi’s schemes with an adaptation of Jacob- Figure 1: Running example graph. son’s index to make it suitable for compressing NULL values and empty adjacency lists in GDBMSs. Our final structure (iii) edge properties. In every native GDBMS we are aware allows constant-time access to NULL or non-NULL values of, the topology is stored in data structures that organize with a small increase in storage overhead per entry compared data in adjacency lists [28], such as in compressed sparse to prior techniques. row (CSR) format. Typically, given the ID of a vertex v, List-based Processing: In Section6, we observe that tradi- the system can in constant-time access v’s adjacency list, tional Volcano-style [38] tuple-at-a-time processors, which which contains a list of (edge ID, neighbour ID) pairs. Typ- are used in some GDBMSs, do not benefit from processing ically, an adjacency list of v is further clustered by the blocks of data in tight loops but also has the advantage of edge label which enables traversing the neighbourhood of avoiding expensive data copies when performing many-to- v based on a particular label efficiently. Vertex and edge many joins, e.g. in long path queries. On the other hand, properties can be stored in a number of ways. For exam- columnar RDBMSs have block-based [20, 64] query proces- ple, some systems use a separate key-value store, such as sors that process fixed-length blocks of data in tight loops, DGraph [2] and JanusGraph [5], and some use a variant of which achieves better CPU and cache utility. However, interpreted attribute layout [25], where records consist of se- block-based processing results in expensive data copies un- rialized variable-sized key-value properties. These records der many-to-many joins. To address this, we propose a new can be located consecutively in disk or memory or have block-based processor we call list-based processor (LBP), pointers to each other, as in Neo4j. which modifies traditional block-based processors in two ways Queries in GDBMSs consist of a subgraph pattern Q that to tailor them for GDBMSs: (i) Instead of representing the describes the joins in the query (similar to SQL’s FROM) intermediate tuples processed by operators as a single group and optionally predicates on these patterns with final group- of equal-sized blocks, we represent them as multiple groups by-and-aggregation operations. We assume a GDBMS with of blocks. We call these list groups. LBP avoids expensive a query processor that uses variants of the following rela- data copies by flattening blocks of some groups into sin- tional operators, as in many GDBMSs, e.g., Neo4j [9], Mem- gle values when performing many-to-many joins. Flattening graph [6], or GraphflowDB: does not require any data copies and happens by setting an SCAN: Scans a set of vertices from the graph. index field of a list group. (ii) Instead of fixed-length blocks, JOIN (e.g. EXPAND in Neo4j and Memgraph, EXTEND in Graph- LBP uses variable length blocks that take the lengths of ad- flowDB): Performs an index nested loop join using the ad- jacency lists that are represented in the intermediate tuples. jacency list index to match an edge of Q. Takes as input a Because adjacency lists are already stored in memory con- partial match t that has matched k of the query edges in Q. secutively, this allows us to avoid materializing adjacency For each t, extends t by matching an unmatched query edge lists during join processing, improving query performance. qvs→qvd, where one of qvs or qvd has already been matched. We integrated our techniques into GraphflowDB [44]. We For example if qvs has already been matched to data ver- present extensive experiments that demonstrate the scala- tex vi, then the operator reads the forward adjacency list bility and performance benefits (and tradeoffs) of our tech- {(ei1, vi1), ..., (ei`, vi`)} of vi to produce a (k+1)-match that 1 niques both on microbenchmarks and on the LDBC social matches one edge-neighbor pair (eij , vij ) of vi . networking benchmark against a row-based Volcano-style FILTER: Applies a predicate ρ to a partial match t, reading implementation of the system, the open-source version of a any necessary vertex and edge properties from storage. commercial GDBMSs, and a column-oriented RDBMS. Our GROUP BY AND AGGREGATE: Performs a standard group by code, queries, and data are available here [15]. and aggregation computation on a partial match t. Example 1. Below is an example query written in the Cypher language [35]: 2. BACKGROUND We review the general storage components and primary MATCH (a:PERSON)−[e:WORKAT]→(b:ORG) query plan operators of GDBMSs. In the property graph WHERE a.age > 22 AND b.estd < 2015 RETURN* model, vertices and edges have labels and arbitrary key value properties. Figure1 shows a property graph that The query returns all persons a and their workplaces b, where we will use as our running example. The graph contains a is older than 22 and b was established before 2014. Fig- ure2 shows a typical plan for this query. vertices with PERSON and ORGANIZATION (ORG) labels, and edges with FOLLOWS, STUDYAT and WORKAT labels. There are three storage components of GDBMSs: (i) topol- 1GraphflowDB can perform an intersection of multiple ad- ogy, i.e., adjacencies of vertices; (ii) vertex properties; and jacency lists if the pattern is cyclic (see reference [50]).

2 SCAN FILTER JOIN FILTER Data Columnar data structure a a.age > 22 a b b.estd < 2015 Vertex Properties V-Column V-Column: of src when n-1, of dst when 1-n, of either src or dst when 1-1 Figure 2: Query plan for the query in Example1. Edge Properties Single-indexed prop. pages when n-n 3. GUIDELINES AND DESIDERATA Fwd Adj. lists V-Column when 1-1 and n-1, CSR o.w. We next outline a set of guidelines and desiderata for or- Bwd Adj. lists V-Column when 1-1 and 1-n, CSR o.w. ganizing the physical data layout and query processor of GDBMSs. We assume edges are doubly-indexed in both a Table 1: Columnar data structures and data components forward and a backward adjacency list, which holds in all they are used for. V-Column stands for vertex column. GDBMSs we are aware of. We will not optimize to avoid this duplication because this is fundamental for performing FOLLOWS STUDYAT WORKAT fast joins on source or destination vertices of edges. p1 p2 p3 ... p1 p2 p3 ... p1 p2 p3 ... Guideline 1. Edge and vertex properties are read in the same order as the edges appear in adjacency lists after joins. e1,p2 e9,p4 e7,p3 e11,p4 e2,p1 e5,p2 e3,p4 ... e4,o1 e12,o2 e8,o2 ... e10,o2 ... Observe that JOIN accesses the edges and neighbors of a vertex v in the order these edges appear in v ’s adjacency i i Figure 3: Example forward adjacency lists implemented as list L = {(e , v )..., (e , v )}. If the next operator also vi i1 i1 i` i` a 2-level CSR structure for the example graph. needs to access the properties of these edges or vertices, e.g., FILTER in Figure2, these accesses will be in the same order. (ii) Label determines properties on vertices and edges. Sim- Our first desiradata is to store the properties of ei1 to ei` sequentially in the same order. Ideally, a system should ilar to the attributes of a relational table, properties on an edge or vertex and their datatypes can sometimes be also store the properties vij sequentially in the same order but in general this would require prohibitive data replication determined by the label. For example, this is the case for every vertex and edge label in LDBC. because while each eij appears in two adjacency lists, each vij appears in as many lists as the degree of vij . (iii) Edges with single cardinality. Edges might have car- Desideratum 1. Store and access the properties of edges dinality constraints. We will differentiate between the sequentially in the order edges appear in adjacency lists. following cardinality constraints: one-to-many (single cardinality in the backward edges), many-to-one (sin- Guideline 2. Access to vertex properties will not be to gle cardinality in the forward edges), one-to-one, and sequential locations and many adjacency lists are very small. many-to-many. An example of one-to-many cardinal- Guideline1 implies that we should expect random accesses ity from LDBC SNB is that each organization has one in memory when an operators access vertex properties. In isLocatedIn edge. addition, real-world graph data with many-to-many rela- tionships have power-law degree distributions, so often there We refer to edges that satisfy properties (i) and (ii) as struc- are many very short adjacency lists in the dataset. There- tured edges and properties that satisfy property (ii) as struc- fore when processing queries with two or more joins, read- tured vertex/edge property. Other edges and properties will ing different adjacency lists will require iteratively reading be called unstructured. The existence of such structure in a short list followed by a random access. This has an im- some graph data motivates our third desideratum: portant implication for adopting compression techniques in Desideratum 3. Exploit structure in the data for space- in-memory GDBMSs. Specifically, techniques that require efficient storage and faster access to data. decompressing blocks of data, say a few KBs, to only read a single vertex property or a single short adjacency list can be prohibitively expensive. Our second desideratum is: 4. COLUMNAR STORAGE Desideratum 2. If compression is used, decompressing We next explore using columnar structures for storing arbitrary data elements in a compressed block should be pos- data in GDBMSs to meet the desiderata from Section3. sible in constant time. For reference, Table1 presents the summary of the colum- nar structures we use and the data they store. Some tech- Guideline 3. Graph data often has partial structure. niques we review here are not novel and even widely used in Although the property graph data model is semi-structured, systems, such as CSR for storing edges. We review them to data stored in GDBMSs often have some structure, which present a complete columnar storage design for a GDBMS. GDBMSs can exploit. In fact, several vendors from industry Others are either new designs, such as single-indexed prop- and academics are actively working on defining a schema erty pages and its accompanying edge ID scheme, or new ap- language for the property graph data model [29, 40]. We plications of existing designs, such as using vertex columns identify three types of structure in property graph data: to store single cardinality edges. (i) Edge label determines source and destination vertex la- bels. Often, edges with a particular label are between a 4.1 CSR for n-n Edges fixed source and destination vertex label. For example, CSR is an existing columnar structure that is widely used in the popular LDBC benchmark (SNB), by existing GDBMSs to store edges. A CSR, shown in Fig- KNOWS edges exist only between vertices of label PERSON. ure3, is a columnar structure that effectively stores a set

3 of (vertex ID, edge ID, neighbour ID) triples sorted by ver- e1(p1,p2) e7(p2,p3) e9(p1,p4) e11(p2,p4) tex ID, where the vertex IDs are compressed similar to run- length encoding. The main difference is that instead of stor- p1 2015 2012 2011 1992 ing vertex ID-run length pairs, CSR stores a list of prefix- e3(p3,p4) e13(p4,p2) e5(p3,p2) e2(p3,p1) sums of the run lengths in the vertex ID column. These 2009 2006 prefix sums are used to directly access the adjacency list of p2 2003 1999 a vertex. In this work, we store the edges of each edge label since INT with n-n cardinality in a separate CSR. As we discuss in Section 4.3, we can store the edges with other cardinalities Figure 4: Single-indexed property pages for since property more efficiently than a CSR using vertex columns. of FOLLOWS edges in the example graph. k = 2.

4.2 Single-indexed Edge Property Pages for need to first find the offset o of e1 in Lnbr1 , which may re-

Properties of n-n Edges quire scanning the entire adjacency list Lnbr1 of nbr1. This Recall Desideratum1 that access to edge properties should access is not constant time. be in the same order of the edges in adjacency lists. We first To overcome this, we can adopt a new edge ID scheme review two columnar structures, edge columns and double- that stores the following: (edge label, source vertex ID, list- 2 indexed property CSRs, the former of which has low storage level positional offset) . With this scheme a system can: (i) cost but does not satisfy Desideratum1 and the latter has identify each edge, e.g., perform equality checks between two high storage cost but satisfies Desideratum1. We then de- edges; and (ii) read the offset o directly from edge IDs, so scribe a new design, which we call single-indexed property reading edge properties in the opposite direction (backward pages, which has low storage cost as edge columns and with in our example) can now be constant time. In addition, this a new edge ID scheme can partially satisfy Desideratum1, scheme can be more space-efficient than schemes that assign so dominates edge columns in this design space. consecutive IDs to all edges as the its first two components Edge Columns: We can use a separate edge column for can often be compressed (see Section 5.2). However, single- indexed property CSR and this edge ID scheme has two each property qi,j of edge label lei. Then with an appro- priate edge ID scheme, such as (edge label, label-level posi- limitations. First access to properties in the ‘opposite direc- tional offset), one can perform a random access to read the tion’ requires two random accesses, e.g., first access obtains the P list using nbr ’s ID and the second access reads a qi,j property of an edge e. This design has low storage cost nbr1 1 and stores each property once but does not satisfy Desider- qi,j property from Pnbr1 . Second, although we do not focus atum1, as it does not necessarily store the properties of the on updates in this paper, using edge IDs that contain po- edges according to their appearance in the adjacency lists. sitional offsets has an important consequence for GDBMSs. In practice, the order would be determined by the sequence Observe that positional offsets that are used by GDBMSs of edge insertions and deletions. are explicitly stored in data structures. For example, in ev- Double-Indexed Property CSRs. An alternative is to ery native GDBMS we are aware of vertex IDs are used as mimic the storage of adjacency lists in the CSRs in separate positional offsets to read vertex properties and they are also CSRs that store edge properties. For each vertex v we can explicitly stored in adjacency lists. This is unlike traditional columnar RDBMSs, where positional offsets, specifically row store qi,j twice in forward and backward property lists. This design provides sequential read of properties in both direc- IDs of tuples, are implicit and not explicitly stored. There- tions, thereby satisfying Desideratum1, but also requires fore, when deletions happen, GDBMSs initially leave gaps double the storage of edge columns (plus the CSR offset in their data structures, and recycle deleted IDs when in- overheads). This can often be prohibitive as many graphs sertions happen. For example, Neo4j’s nodestore.db.id file have orders of magnitude edges than vertices and avoiding keeps track of deleted IDs for later recycling [8]. Similarly, duplicate storage of large properties, such as strings, is de- the list-level positional offsets need to be recycled. This may sirable especially for in-memory systems, which we focus on. leave many gaps in adjacency lists because to recycle a list- A natural question is: Can we avoid the duplicate stor- level offset, the system needs to wait for another insertion age of double-indexed property CSRs but at the same time into the same adjacency list, which may be infrequent. achieve sequential reads? We next show a structure that Our single-indexed property pages addresses these two is- with an appropriate edge ID scheme obtains sequential reads sues (Figure4). We store k property lists (by default 128) in one direction, so partially satisfying Desideratum1. This in a property page. Within a property page, properties of structure therefore dominates edge columns as a design. the same list are not necessarily stored consecutively. How- Single-indexed property pages: A first natural design ever, because we use a small value of k, these properties uses only one property CSR, say forward. We call this struc- are stored in close-by memory locations. We modify the ture single-indexed property CSR. Then, properties can be edge ID scheme above to use page-level positional offsets. read sequentially in the forward direction. However, reading This has two advantages. First, given a positional offset, a property in the other direction quickly, specifically with the system can directly read an edge property (so we avoid constant time access, requires a new edge ID scheme. the access to read Pnbr1 ). Second, the system can recycle a To see this suppose a system has read the backward ad- page-level offset whenever any one of the k lists get a new insertion. Figure4 shows the single-indexed property pages jacency lists of a vertex v with label lei, {(e1, nbr1), ..., in the forward direction for property of edges with (ek, nbrk)}, and needs to read the qi,j property of these since label when k=2 for reference. edges. Then given say e1, we need to be able to read e1’s qi,j FOLLOWS property from the forward property list Pnbr1 of nbr1. With a standard edge ID scheme, for example one that assigns 2If we use the backward property CSR, the second compo- consecutive IDs to all edges with label lei, the system would nent would instead be the destination vertex ID.

4 vertex properties STUDYAT n-1 WORKAT n-1 pression completely or support decompressing arbitrary el- ements in a block in constant time. This is only possible if the elements are encoded in fixed-length codes instead of 0 peter 17 M p1 e4 2019 variable-length codes. We review dictionary encoding and

1 alice 45 F p2 e12 2014 leading 0 suppression, which we integrated in our implemen-

2 jenny 23 F p3 e8 2006 e10 2006 tation and refer readers to references [19, 36, 47] for details of other fixed-length schemes, such as frame of reference. 3 bob 54 M p4 e6 1980

Dictionary encoding: This is perhaps the most common ...... encoding scheme to be used in RDBMSs [19, 66]. This gender name age STUDYAT doj STUDYAT doj scheme maps a domain of values into compact codes using STRING INT STRING EDGE INT EDGE INT a variety of schemes [19, 39, 66], some producing variable- length codes, such as Huffmann encoding, and others fixed- Figure 5: Example vertex columns storing vertex length codes [19]. We use dictionary encoding to map a properties and single-cardinality edges and their properties. categorical edge or vertex property p, e.g., gender property of PERSON vertices in LDBC SNB dataset, that takes on z 4.3 VertexColumns for VertexProperties, Sin- different values to dlog2(z)/8e bytes (we pad log2(z) bits gle Cardinality Edges and Edge Properties with 0s to have a fixed number of bytes). Leading 0 Suppression3: Given a block of data, this With an appropriate vertex ID scheme, columns can be scheme omits storing leading zero bits in each value of the directly used for storing structured vertex properties in a block [25]. We adopt a fixed-length variant of this for storing compact manner. Let pi,1, pi,2, ...pi,n be the structured ver- components of edge and vertex IDs, e.g., if the maximum size tex properties of vertices with label lvi. We can store each of a property page of an edge label is t, we use dlog2(k)/8e pi,j in a vertex column, that store pi,j properties of vertices many bytes for the page-level positional offset of edge IDs. in consecutive locations. Then we can adopt a (vertex la- bel, label-level positional offset) ID scheme and ensure that 5.2 Factoring Out Edge/VertexID Components offsets with the same label are consecutive. Similar to the Our vertex and edge ID schemes from Sections4 decom- edge ID scheme from Section 4.2, this ID scheme also can be pose the IDs into multiple small components, which enables compressed by factoring out vertex labels (see Section 5.2). us to factor out most of these components, when the data We next observe that we can exploit single cardinality depicts some structure (Desideratum3). This enables us to constraints, i.e., 1-1, 1-n, or n-1 constraints, on the edges to compress the adjacency lists without needing to decompress store them and their properties more efficiently than CSR while scanning them. Recall that the ID of an edge e is a or property pages (Desideratum3). Instead of a CSR or a triple (edge label, source/destination vertex ID, page-level property page, we can use vertex columns. Specifically, we positional offset) and the ID of a vertex v is a pair (vertex la- can store these edges and their properties directly as a prop- bel, label-level positional offset). Recall also that GDBMSs erty of source or destination vertex of the edges and directly store (edge ID, neighbour ID) pairs inside adjacency lists. access them using a vertex positional offset. This is more First, the vertex IDs inside the edge ID can be omitted be- efficient than a 2-level CSR, or property pages which require cause this is the neighbour vertex ID, which is already stored arithmetic on both the source or destination ID and a page- in the pairs in the lists. Second edge labels can be omitted level offset to access properties. As we discuss in the next because we cluster our adjacency lists by edge label. The section, by storing the properties of single cardinality edges only components of (edge ID, neighbour ID) pairs that need in vertex columns, we can also compress their edge IDs fur- to be stored are: (i) positional offset of the edge ID; and ther because we no longer need page-level positional offsets. (ii) vertex label and positional offset of neighbour vertex Figure5 shows single cardinality STUDYAT and WORKAT ID. When the data depicts some structure, we can further edges from our example and their properties stored as vertex factor out some of these components as follows: column of PERSON vertices. • Edges do not have properties: Often, the edges of a par- ticular label do not have any structured or unstructured 5. COLUMNAR COMPRESSION properties. They only represent the existence of relation- Data compression and query processing directly on com- ships between vertices. For example 10 out of 15 edge pressed data are extensively used techniques in columnar labels in LDBC SNB do not have any properties. When RDBMSs. To present our complete integration, we start by edges do not have properties, they do not need to be iden- reviewing techniques that are directly applicable to GDBMSs tifiable, as the system will not access their properties. and not novel. We then discuss the cases when we can com- Therefore, we can distinguish two edges by their neigh- press the new vertex and edge ID schemes from Section4. bour vertex ID and edges with the same IDs are simply Finally, we review existing NULL compression schemes from replicas of each other. Hence, we can completely omit columnar RDBMSs [17, 19] and show their shortcomings storing the positional offsets of edge IDs. and enhance one of them with Jacobson’s bit vector index • Edge label determines neighbour vertex label. Often, edge to make the suitable for GDBMSs. labels in the graph are between a single source and des- tination vertex label. In this case, we can omit storing 5.1 Directly Applicable Techniques the vertex label of the neighbour ID. For example, this Recall our Desideratum2 that because access to vertex property exists for 10 edge labels in LDBC SNB. properties cannot be localized and because many adjacency 3In reference [19], Abadi et al. refer to this as NULL sup- lists are very short, the compression schemes that are suit- pression, not to be confused with NULL compression to com- able for in-memory GDBMSs need to either avoid decom- presses actual NULL values in columns (see Section 5.3).

5 edge e 7 6 3 2 8 11 2 3 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 Uncompressed block

single no has yes notNull values 7 6 3 2 8 11 2 3 5 cardinality ? properties ? bitStrings 0100b 0111b 0110b 0011b

yes no prefix sums 0 1 4 6 Do not store Store positional positional offsets offsets Prefix sum-based NULL Compressed block 0 1 2 3

Figure 6: Decision for storing page-level positional ...... offsets of edges in adjacency lists. finding value at 9 : 0101b 0 1 1 2 chunkIdx = 2 0110b 0 0 1 2 • Single cardinality edges: Recall from Section 4.3 that the position of element in chunk (i) = 1 0111b 0 1 2 3 properties for single cardinality edges can be stored in position in notNull values array = 4 + 0 = 4 ...... vertex columns. So we can directly read these properties by using the source or destination vertex ID. Equivalently Bit position to index Map the page-level positional offsets of a single cardinality edge Figure 7: NULL compression using a simplified Jacobson’s is always 0, so can be omitted. For example, this property bit vector rank index with chunk size 4. exists in 8 of the edge labels in LDBC SNB. Figures6 shows our decision tree to decide when to omit storing the page-level positional offsets in edge IDs. and the bit-string, we store prefix sums for each c (16 by default) elements in a block of a column, i.e., we divide the 5.3 NULL and Empty List Compression block into chunks of size c. The prefix sum holds the number Edge and vertex properties can often be quite sparse in of non-NULL elements before the current chunk. We also maintain a pre-populated static 2D bit-string-position-count real-world structured graph data. Similarly, due to the power- c law nature of degree distributions, many vertices can have map M with 2 cells. M[b, i] is the number of 1s before the empty adjacency lists in CSRs. Both can be seen as different i’th bit of a c-length bit string b. Let p be the offset which columnar structures containing NULL values, which can be is non-NULL and b the c-length bit string chunk in the bit compressed. vector that p belongs to. Then rank(p) = ps[p/c] + M[b, p A naive way to compress NULLs in columnar RDBMSs mod c]. This is a simplified version of Jacobson’s original in- is to treat NULL values as a separate data value and use dex, which divides the bit vector into two sets of prefix sums. run-length encoding [17]. Abadi in reference [17] describes This slightly reduces the space for storing prefix sums but a design space of more optimized techniques for compress- also requires twice as many operations to compute ranks. ing NULLs in columns. All of these techniques list non- The choice of c primarily affects how big the pre-populated NULL elements consecutively in a ‘non-NULL values col- map is. A second important parameter in this scheme is the umn’ and use a secondary structure to indicate the positions number of bits m that we use for each prefix sum value, of these non-NULL elements. First technique lists positions which determines how large a block we are compressing and how much overhead the scheme has for each element. For an of each non-NULL value consecutively, which is suitable for c very sparse columns, e.g., with > 90% NULLs. Second, for arbitrary m, c, we require: (i) 2 ∗ log(c) size map, because the map has 2c cells and needs to store a log(c)-bit count dense columns, lists non-NULL values as a sequence of pairs, m each indicating a range of positions with non-NULL values. value in each cell; (ii) we can compress a block of size 2 ; and Third, for columns with intermediate sparsity, uses a bit (iii) we store one prefix sum for each c elements, so incur a vector to indicate if each location is NULL or not. The last cost of m/c extra bits per element. By choosing m = 16, c = 16 (so we use short integers), we require 2c ∗ c = 1MB-size technique is quite compact and requires only 1 extra bit per m each element in a column. map, can compress 2 = 64K blocks, and incur m/c = 1 However, none of these techniques are directly applica- extra bit overhead for each element, so increase the overhead ble to GDBMSs because they do not allow constant-time of reference [17]’s scheme from 1 to only 2 bits per element access to non-NULL values, so do not satisfy our Desider- (but provide constant time access to non-NULL values). atum2. Instead, these schemes are designed assuming an operator will read in tandem an entire block of non-NULL 6. LIST-BASED PROCESSING values column and secondary data structure sequentially. We next describe our list-based query processor (LBP). To support constant-time access to non-NULL values, the We first motivate LBP by discussing the shortcomings of tra- secondary structure needs to support two operations in con- ditional Volcano-style tuple-at-a-time query processors and stant time: (i) check if it is NULL or not; and (ii) if it is the traditional block-based processors of columnar RDBMSs non-NULL, we need to compute the rank of p, i.e., compute when processing many-to-many joins. We give an example. the number of non-NULL values before p. Abadi’s third design that uses a bit vector already sup- Example 2. Consider the following query. P, F, S, and ports checking if an arbitrary position p is NULL. To support O abbreviate PERSON, FOLLOWS, STUDYAT, and ORGANISATION. rank queries, we enhance this design with a simplified ver- MATCH (a:P)−[:F]→(b:P)−[:F]→(c:P)−[:S]→(d:O) sion of Jacobson’s bit vector index [42, 43]. Figure7 shows WHERE a.age > 50 and d.name = "UW" RETURN* this design. In addition to the array of non-NULL values

6 List Group 1 List Group 2 List Group 3 SCAN FILTER JOIN jo1 JOIN jo2 JOIN jo3 FILTER a.age d.name c.ID c c ... c a.ID a1 a2 ... a1024 b.ID b1 b2 ... bk1 k2+1 k2+2 2k2 a > 50 a b b c c d = “UW” filter a.age 51 19 ... 23 1 1 ... 1 d.ID d d ... d mask k2+1 k2+1 2k2 filter 1 0 ... 0 d.name UW UofT ... UW mask curIdx = 2 Figure 8: Query plan for the query in Example2. filter 1 0 ... 1 curIdx = 1 mask curIdx = -1 Consider evaluating a simple plan for this query shown in Figure8, which is akin to a left-deep plan in an RDBMS, Figure 9: Intermediate chunk for the query in Example2. on a graph where FOLLOWS are many-to-many edges and The first two list groups are flattened to single tuples, STUDYAT edges have single cardinality. Volcano-style tuple- while the last one represents k2 many tuples. at-a-time processing [38], which some GDBMSs adopt [9, 50], is efficient in terms of how much data is copied to the example we had three variables a, b, and c corresponding intermediate tuple. Suppose the scan matches a to a1 and to three arrays. The values at position i of all arrays form a1 extends to k1 many b’s, b1 . . . bk1 , and each bi extends a single tuple. Therefore to represent the tuples that are to k2 many c’s to bik2 ..., b(i+1)k2 (let us ignore the d exten- produced by many-to-many joins, repetitions of values are sion for now). Even though this generates k1 × k2 tuples, necessary. One approach to address this can be to compress the value a1 would be copied only once to the tuple. This the arrays that correspond to the variables that will be re- is an important advantage for GDBMSs which frequently peated with run-length encoding. This however incurs the process many-to-many joins. However, it is well known that overhead of updating the run length values of each of these Volcano-style processors do not achieve high CPU and cache variables each time a join extends partial matches. Instead utility as processing is intermixed with many iterator calls. LBP breaks the arrays into multiple groups of blocks, which Column-oriented RDBMSs instead adopt block-based pro- we call list groups, which we explain in detail momentarily. cessors [26, 41], which processes an entire block at a time Second, instead of using fixed-length blocks, the blocks in in operators4. Block sizes are fixed length, e.g. 1024 tu- each group can take different lengths, which are aligned to ples [3, 27]. While processing blocks of tuples, operators the lengths of adjacency lists in the database. As we shortly read consecutive memory locations, achieving good cache explain, this allows us to avoid materializing adjacency lists locality, and perform computations inside loops over ar- into the blocks. Specifically, the size of a list group is either rays which is efficient on modern CPUs. However, tra- fixed, by default 1024, if the group has a block that contains ditional block-based processors have two shortcomings for vertices and vertex properties that are output by an initial GDBMSs. (1) For many-to-many joins, block-based pro- scan operator. Otherwise, the output vertices, edges and cessing requires more data copying into intermediate data properties of one-to-many or many-to-many join operators structures than tuple-at-a-time processing. Suppose for sim- are put into new list groups whose sizes are dynamic and plicity a block size of k1 ×k2. In our example, the scan would take different adjacency list sizes during runtime. If a one- output an array a :[a1], the first join would output a : to-one or many-to-one join operator extends say vertices u to

[a1, ..., a1], b :[b1, ..., bk1 ] vectors, and the second join would v, then the blocks that contain vertices v, the u→v edges and output a :[a1, ..., a1], b :[b1, ..., b1, b2, ..., b2, ...., bk1 , ..., bk1 ], their properties are put into an existing list group. We call c :[c1, ..., ck2 , ..., c(k1−1)k2 , ...,ck1k2 ], where for example the the union of list groups intermediate chunk, which represents value a1 gets copied k1 × k2 times into intermediate arrays. a set of intermediate tuples. Instead of copying data values, some column-oriented sys- Example 3. Continuing our example, suppose that each tems, such as DuckDB [3], keep a small ‘data array’ with cj has a (single) STUDYAT edge to dj . Figure9 shows the a single a1 value, and use additional ‘selection arrays’ that three list groups when LBP evaluates our running example contain copies of offsets that point to the data values (so query using the same plan in Figure8. The first group has instead offset 0 is now copied many times). (2) Traditional the a.ID and a.age blocks of size 1024, output by the Scan block-based processors do not exploit the list-based data or- operator. The second and third groups have blocks of size ganization of GDBMSs. Specifically, the adjacency lists that k1 and k2. The sizes of these groups are determined dur- are used by join operators are already stored consecutively ing run time by the join operators jo1 and jo2 that respec- in memory, which can be exploited to avoid materializing tively extend vertices a1 and b2 to their k1 and k2 many these lists into blocks. FOLLOWS edges. The third join operator, which extends each We developed a new block-based processor, which we call c vertex to one d vertex through a single cardinality STUDYAT list-based processor (LBP) that addresses these shortcom- edge, does not create a new group. Instead, d.ID and d.city ings. LBP has two differences from traditional block-based blocks are also part of the third list group. processors, each of which addresses one of the two short- comings we identified above. First, traditional block-based Each list group has another field curIdx, which can ei- processors represent a set of tuples as a single group of ther be: (i) a non positive integer that is less than the block blocks/arrays, each corresponding to one variable. In our sizes in the group; or (ii) -1. If curIdx ≥ 0, the list group is flattened and represents a single tuple that consists of 4Block-based processing has been called vectorized process- the curIdx’th values in the blocks. For example, as shown ing in the original work on MonetDB/X100 from CWI [65]. with vertical rectangles in Figure9, the first and second We use the term block-based not to confuse it with SIMD list groups are flattened to their first and second values, re- vectorized instructions. For example, a new column store DuckDB [3] also from CWI uses block-based processing but spectively. Otherwise, the list group is not flattened and without SIMD instructions. Note however that operators in represents as many tuples as the size of its blocks. Each block-based processors can use SIMD instructions as they list group also has an optional filter mask that identifies process multiple tuples at a time inside loops. the tuples that are actually in the block and have not been

7 filtered out. The set of tuples that the entire intermediate then perform a detailed evaluation of each of our optimiza- chunk represents are the Cartesian product of the tuples tions using microbenchmark queries on several datasets. We from each list group. Therefore, this is a form of factoriza- end with an end-to-end baseline performance comparisons of tion [55], and is more lightweight than run-length encoding GF-CL against GF-RV, the commercial Neo4j GDBMS system, of blocks to compress intermediate results. and the Vertica and MonetDB columnar RDBMS systems Importantly, similar to traditional block-based processors, on the LDBC social network benchmark (SNB). Unless oth- every primitive data operator in LBP performs computa- erwise specified, we run queries on GF-CL. tions on blocks inside loops. For example, the first filter operator in Figure8 runs on the 1024 tuples in the first list 7.1 Experimental Setup group to filter those with age > 50 and write the necessary Hardware Setup: For all our experiments, we use a single bits to the filter mask field. Join operators that perform machine that has two Intel E5-2670 @2.6GHz CPUs and one-to-many or many-to-many joins are the only operators 512 GB of RAM. We only use one logical core. We set the that flatten list groups. For example consider jo1 in Figure8 maximum size of the JVM heap to 500 GB and keep JVM’s that extends a vertices to their possibly many b neighbors. default minimum size. This operator loops through the 1024 values in the a block, Dataset: We generated LDBC [33] benchmark’s social net- iteratively flattens the first list group by setting its curIdx work data using scale factors 10 and 100, which we refer field, and then extends each a value to a list of b values, to as LDBC10 and LDBC100, respectively. In LDBC, all which are stored in the second list group. So subsequent of the edges and edge and vertex properties are structured operators get a flattened list group 1 and an unflattened list but several properties and edges are very sparse. LDBC10 group 2. One-to-one or many-to-one join operators, e.g., jo3 contains 30M vertices and 176.6M edges while LDBC100 in Figure8, do not flatten list groups because they take a contains 1.77B edges and 300M vertices. Both datasets con- block of values, e.g. c, and produce new blocks with the tain 8 vertex labels, 15 edge labels, 29 vertex properties, and same length, e.g., d. Representing intermediate tuples as 5 edge properties. We use LDBC datasets in both our end- a set of list groups also allows LBP to perform fast group to-end and and micro benchmark experiments. To enhance by and aggregations. For example, our count(*) aggregator our microbenchmarks further, we use two datasets from the simply multiplies the sizes of each list group to compute the popular Konect graph sets [46] covering two application do- number of tuples represented by each intermediate chunk mains: a Flickr social network (FLICKR)[51] and a Wikipedia it receives. This is similar to columnar RDBMSs and fac- hyperlink between articles of the German Wikipedia graph torized query processors [23, 24] that perform computations (WIKI)[16]. Flickr graph has 2.3M nodes and 33.1M edges directly on compressed data [19, 66]. while Wikipedia graph has 2.1M nodes and 86.3M edges. Finally, LBP avoids materializing two types of blocks to Both datasets have timestamps as edge properties. improve performance. These are shown in gray shade in Fig- We describe our queries within each sub-section. The ex- ure9. First are vertex IDs from initial vertex scans, in our act queries used in each experiment can be found in our example a.ID, which are simply consecutive integers. We repo [15]. In each experiment, we ran our queries 5 times have a special block implementation, called nodeIDSequence, consecutively and report the average of the last 3 runs. that supports our block interface and avoids any data copies. Second, and more importantly, results of one-to-many and 7.2 Memory Reduction many-to-many joins, which are lists of edge IDs and neigh- We demonstrate the memory reduction we get from the bor vertex IDs are also not materialized. These ID lists columnar storage and compression techniques we studied us- are already stored consecutively in CSR structures, so we ing LDBC100. We start with GF-RV and integrate one addi- can exploit this design and implement these blocks simply tional storage optimization step-by-step ending with GF-CL: as pointers to parts of CSRs. This can allow us to perform some computations with very little or even without any data (i) +COLS: Stores vertex properties in vertex columns, edge copying into the intermediate chunk. For example, queries properties in single-directional property pages, and sin- that counts k-degree neighbors of a vertex, e.g., MATCH gle cardinality edges in vertex columns (instead of CSR). (a)→(b)→(c) WHERE a.ID=0 RETURN count(*), are eval- (ii) : Introduces our new vertex and edge ID schemes uated in plans where operators either use a nodeIDSequence +NEW-IDS block or blocks that point to existing adjacency lists, so do and factors out possible ID components (recall Section 5.2). not copy any data into the intermediate chunk. (iii) +0-SUPR: Implements leading 0 suppression in the com- ponents of vertex and edge IDs in adjacency lists. 7. EVALUATIONS (iv) +NULL: Implements NULL compression of empty lists We integrated our columnar techniques into GraphflowDB, and vertex properties based on Jacobson’s index. which is an in-memory GDBMS written in Java. We refer to this version of GraphflowDB as GF-CL for Columnar List- Table2 shows how much memory each component of the based. We based our work on the publicly available version system as well as the entire system take after each opti- here [4], which we will refer to as GF-RV, for Row-oriented mization. We see 2.96x and 2.74x reduction for storing for- Volcano. GF-RV uses 8 byte vertex and edge IDs and adopts ward and backward adjacency lists, respectively. We reduce the interpreted attribute layout to store edge and vertex memory significantly by adopting our new ID scheme and properties. GF-RV also partitions adjacency lists by edge la- factoring out components, such as edge and vertex labels, bels and stores the edge ID-neighbor ID pairs inside a CSR. and using small size integers for positional offsets. We also We first present an experiment evaluating the memory im- see 1.58x reduction by storing vertex properties in columns, provements of GF-CL over GF-RV on a large social network which avoids the overhead of storing the keys of the proper- dataset generated by the popular LDBC benchmark. We ties explicitly, which the interpreted attribute layout design

8 GF-RV +COLS +NEW-IDS +0-SUPR +NULL GF-CL edges. For LDBC the paths enumerate Knows edges (WIKI Vertex 31.40 19.87 19.87 19.87 19.28 - and FLICKR contain only one edge label). Our 1-hop query Props. +1.58x - - +1.03x 1.62x compares the edge’s timestamp for WIKI and FLICKR and the Edge 7.92 2.07 2.07 2.07 2.05 - creationDate property for LDBC to be greater than a con- Props. +3.82x - - +1.01x 3.87 stant. Our 2-hop query compares the property of each query F. Adj. 31.93 28.95 20.67 11.41 10.78 - edge to be greater than the previous edge’s property. For Lists +1.10x +1.40x +1.81x +1.06x 2.96 WIKI, which contain prohibitively many 2-hops we further B. Adj. 31.29 31.07 24.93 13.10 11.41 - put a predicate on the source and destination nodes to make Lists +1.01x +1.25x +1.90x +1.15x 2.74 the queries finish within a reasonable time. When running Total 102.56 81.97 67.55 46.45 43.54 - queries in both configurations, for each query, we consider (GB) +1.25x +1.21x +1.45x +1.07x 2.36 two plans: (i) the forward plan that matches vertices from left to right in the forward direction; (ii) the backward plans Table 2: LDBC100 memory consumption improvements on that matches vertices from right to left. each data component after applying one more optimization Forward plans perform sequential reads of properties un- on top of the configuration on the left. GF-CL reports the der PROP-PAGES, which achieves good CPU cache locality. final improvements of GF-CL over GF-RV. We therefore expect them to be more performant than back- ward plans under PROP-PAGES as well as the forward plans LDBC100 WIKI FLICKR under EDGE COLS, which both lead to random reads. We 1H 2H 1H 2H 1H 2H also expect backward plans to behave similarly under both COLE 0.6 71.9 4.8 37.7 1.7 878.7 configurations. Table3 shows our results. Observe that PF 0.2 38.3 1.0 17.6 0.8 185.8 forward plans under PROP-PAGES is between 1.8x to 4.8x PAGEP 2.5x 1.8x 4.8x 2.1x 2.1x 4.7x faster than the forward plans under EDGE COLS and are also COLE 1.3 130.7 7.2 108.7 2.4 1090.0 faster than the backward plans under PROP-PAGES. In con- PB 1.3 136.5 6.4 92.9 2.7 1009.6 trast, the differences between the performances of the back- PAGE P 0.96x 1.0x 1.1x 1.2x 0.89x 1.1x ward plans are comparable. This is because neither edge columns nor forward-directional property pages provide any Table 3: Runtime (in secs) of k-hop (KH) queries when locality when accessing properties in the order of backward storing edge properties in forward-directional property adjacency lists. This verifies our claim in Section 4.2 that pages (PAGEP ) vs edge columns (COLE ) on LDBC100, PROP-PAGES is a strictly better columnar design than using Wiki, and Flickr using a Forward Plan (PF ) and a vanilla edge columns. Backward Plan (PB ) as defined in Secion 7.3. 7.4 VertexColumns for Single Cardinality Edges stores. The modest memory gains in +COLS for storing ad- In Section 7.2, we showed the memory gains of storing jacency lists is due to the fact that 8 out of 15 edge labels single cardinality edges in vertex columns. This storage in LDBC SNB are single cardinality and storing them in also improves performance because the system can direct vertex columns is more lightweight than in CSRs, as it does access the edge without an indirection through a CSR. We not store the CSR offsets. We see a reduction of 3.82x when next demonstrate this benefits under two settings: (i) when storing edge properties in single-directional property pages. empty lists (or edges because of single cardinality) are not This is primarily because GF-RV stores a pointer for each NULL compressed; and (ii) when they are NULL compressed. edge, even if the edges with a particular label have no prop- We create 4 configurations of GraphflowDB: erties. GF-CL stores no columns for these edges, so incur no (i) V-COL-UNC: Single cardinality edge label edges are stored overheads and avoids storing the keys of the properties ex- in vertex columns and are not compressed. This is equiv- plicitly. We see modest benefits in NULL compression since alent to +OMIT configuration in Section 7.2. empty adjacency lists are infrequent in LDBC100 and 26 of (ii) CSR-UNC: Single cardinality edge label edges are stored 29 vertex properties and all of the edge properties contain in CSR format and are not compressed. no NULL values. Overall, we obtained a reduction of 2.36x onLDBC100, reducing the memory cost from 102.5GB to (iii) V-COL-C: Null compressed version of V-COL-UNC. This is 43.5GB, which demonstrate the potential benefits of apply- equivalent to +NULL configuration in Section 7.2. ing our techniques to scale in-memory GDBMSs. (iv) CSR-C: Null compressed version of CSR-UNC. 7.3 Single-Directional Property Pages We use the LDBC datasets only because the other datasets We next demonstrate the query performance benefits of do not contain single cardinality edges. We use LDBC100. storing edge properties in single-directional property pages The workload consists of simple 1-, 2-, and 3-hop queries on compared to the baseline edge column design. We use our the replyOf edge between Comment vertices. To ensure that new ID scheme and configure GraphflowDB in two ways: the joins are the dominant operation in these queries, the EDGE COLS: Stores edge properties in an edge column. We queries do not contain any predicates and return as output assume edges get random global IDs, so their properties can the count star aggregation. We evaluate each query with a appear anywhere within this column. plan that performs the joins in the forward direction. PROP PAGES: Edge properties are stored in forward-directional Table4 shows the result of queries under each system pages by combining k=128 adjacency lists’s properties. configuration. We observe up to 1.62x performance gains We use the LDBC100, WIKI, and FLICKR datasets. As our between uncompressed variants of vertex columns and CSR workload, we use 1- and 2-hop queries, i.e., queries that (i.e., V-COL-UNC vs CSR-UNC) and up to 1.49x gains between enumerate all edges and 2-paths, with predicates on the NULL compressed variants (i.e., V-COL-C vs CSR-C). These

9 1-hop 2-hop 3-hop Mem (in MB) Uncompressed J-NULL Vanilla-NULL CSR-UNC 7.03 9.13 9.60 1266.56 4.34 5.80 5.85 839.93 V-COL-UNC 1.62x 1.57x 1.64x 1.51x CSR-C 7.78 10.40 11.23 905.23 5.23 8.28 8.41 478.86 V-COL-C 1.49x 1.26x 1.34x 1.89x

Table 4: Vertex property columns vs. 2-level CSR adjacency lists for storing single cardinality edges: Query runtime is in seconds and memory usage in MBs. results verify that using vertex columns for single-cardinality edges not only saves space, but also improves query per- Figure 10: Query performance and memory consumption formance irrespective of whether or not the edges/lists are when storing a vertex property column as uncompressed, NULL compressed or not. Recall that in Section 7.2, we compressed with Jacobson’s scheme, and the vanilla bit had reported that NULL compression leads to modest mem- string scheme from Abadi, under different density levels. ory reduction when we look at the size reduction of the entire database (by 1.07x). This was because majority of the edges and properties are not sparse in LDBC. However, column is very sparse, accesses are often to NULL elements, for the storage cost of specific edges, we observe major re- which takes one access for reading the bit value of the el- ductions. The last column of Table4 reports the size of ement. When the bit value is 0 iterators return a global storing replyOf edges under each system configuration. In NULL value which is likely to be in the CPU cache. Instead, LDBC100, 50.5% of the replyOf forward adjacency lists are Uncompressed always returns the value in the element’s cell, empty. Observe that NULL compressing these lists lead to which has a higher chance of a CPU cache miss. 1.75x memory reduction when using vertex columns (from 839.93MB vs 478.86MB). We note that the reduction is 7.6 List-based Processor lower, by 1.4x, if we use CSRs because CSRs incur the cost We next present experiments demonstrating the perfor- of storing extra offsets which cannot be compressed if we mance benefits of LBP against a traditional Volcano-style want to maintain constant time access to adjacency lists. tuple-at-a-time processor, which are adopted by some exist- ing systems, such as Neo4j [9] or MemGraph [6]. Recall that 7.5 Null Compression similar tuple-at-a-time Volcano-style processors, LBP does We further demonstrate the memory/performance trade- not duplicate data values under many-to-many joins but has off of our NULL compression scheme using Jacobson’s index. three advantages over traditional tuple-at-a-time processor: We now focus on compressing sparse property columns. We (1) all primitive computations over data happen inside loops take the following query: “MATCH (a:Person)−[e:Likes]→ as in block-based operators; (2) the join operator can avoid (b:Comment) RETURN b.creationDate”. Then we create copies of edge ID-neighbor ID pairs into intermediate tuples, multiple versions of the LDBC100 dataset, in which the exploiting the list-based storage; and (3) we can perform creationDate property of Comment vertices contains differ- group-by and aggregation operations directly on compressed ent percentage of non-NULL values. LDBC100 contains data. We present two separate sets of experiments that 220M Comment vertices, so our column contains 220M en- demonstrate the benefits from these three factors. To en- tries. We then evaluate this query with a simple plan that sure that our experiments only test differences due to query scans a, extends to b, and then a sink operator that reads processing techniques, we integrated our columnar storage b.creationDate. We compare the query performance and and compression techniques into GF-RV (recall that this is the memory cost of storing the creationDate column, when GraphflowDB with row-based storage and Volcano-style pro- it is stored in three different ways: (i) J-NULL compresses the cessor). We call this version GF-CV, for columnar Volcano, column using Jacobson’s bit index with our default config- and compare GF-CL against GF-CV. uration (m=16, c=16); (ii) Vanilla-NULL is the vanilla bit We use LDBC100, Wikipedia, and Flickr datasets. In our string-based scheme from reference [17]; and (ii) Uncompres- first experiment, we take 1-, 2-, and 3-hop queries (as in sed stores the column in an uncompressed format. Section 7.3, we use the Knows edges in LDBC100), where Figure 10 shows the memory usage and performance of our the last edge in the path has a predicate to be greater than query under three different system configurations. Recall a constant (e.g., e.date > c). For both GF-CV and GF-CL, we that under our default configuration J-NULL requires slightly consider the standard plan that scans the left most node, more memory than Vanilla-NULL, 2 bits for each element in- extends right to match the entire path, and a final Filter stead of 1 bit. The memory reduction due to compression is on the date property of the last extended edge. A major high enough to compensate this overhead until 93.7% of the part of the work in these plans happen at the final join and values are non-NULL. As expected the query performance filter operation, therefore these plans allow us to measure of J-NULL is generally slower than Uncompressed, between the performance benefits of performing computations inside 1.19x and 1.51x, but much faster than Vanilla-NULL, which loops and avoiding data copying in joins. Our results are was >20x slower than J-NULL and is therefore omitted in shown in the FILTER rows of Table5. We see that GF-CL Figure 10. Interestingly, when the column is sparse enough outperforms GF-CV between 2.7x and 15.2x. This demon- (containing less than 30% non-NULL values), J-NULL can strates the significant performance benefits of performing even outperform Uncompressed. This is because when the computations inside loops and avoiding copies of adjacency

10 1-hop 2-hop 3-hop GF-CV 24.6 1470.5 40252.4 FILTER 7.7 116.2 2647.3 GF-CL 3.2x 12.7x 15.2x LDBC100 GF-CV 13.4 241.9 6947.3 COUNT(*) 4.2 18.9 357.9 GF-CL 3.2x 12.8x 19.4x GF-CV 32.6 1300.0 14864.0 FILTER 12.2 95.3 1194.7 GF-CL 2.7x 13.7x 12.4x FLICKR GF-CV 35.3 519.2 4162.5 COUNT(*) 16.9 23.4 51.7 GF-CL 2.1x 21.4x 80.6x GF-CV 35.8 4500.2 236930.2 FILTER 11.9 1192.5 20329.3 GF-CL 2.9x 3.8x 11.7x Figure 11: Relative speedup/slowdown of the different WIKI GF-CV 32.7 1745.2 109000.2 systems in comparison to GF-RV on LDBC10. The boxplots COUNT(*) 19.0 27.6 120.4 show the 5th, 25th, 50th, 75th, and 95th percentiles. GF-CL 1.7x 63.2x 905.1x subgraph patterns in the MATCH clause. We modified the Table 5: Runtime (ms) of tuple-at-a-time (GF-CV) and LBP Interactive Complex Reads (IC) and Interactive Short Reads (GF-CL) plans. (IS) queries from LDBC [33] in order to be able to run them. Specifically GraphflowDB does not support variable length lists into intermediate tuples. queries that search for joins between a minimum and maxi- In our second experiment, we demonstrate the benefits mum length, which we set to the maximum length to make of performing fast aggregations over compressed intermedi- them fixed-length instead, and shortest path queries, which ate results. We modify the previous queries by removing we removed from the benchmark. We also removed predi- the predicate and instead add a return value of count(*), so cates that check the existence or non-existence of edges be- these queries simply count the number of 1-, 2-, and 3-edge tween nodes and the ORDER BY clauses. Our final work- paths. We use the same plans as before except we change the load contains variants of 18 of the 21 queries in the IS and last Filter operator with a GroupBy operator. Our results IC benchmarks. Our exact queries are given in AppendixA. are shown in the COUNT(*) rows of Table5. Observe that We used the the community version v4.2 of the Neo4j the improvements are much more significant now, up to close GDBMS [9], the community version 10.0 of Vertica [14] and to three orders of magnitude on Wiki (by 905.1x). The pri- MonetDB 5 server 11.37.11 [7]. We note that our experi- mary advantage of GF-CL is now that the counting happens ments should not be interpreted as one system being gen- on compressed intermediate results. In addition it is inter- erally more efficient than another. It is very difficult to esting to note that GF-CL does not copy any data into the meaningfully compare completely separate systems due of intermediate chunk for these queries. For example, consider many differences across systems, e.g., all baseline systems the plan for the 2-hop count query. The scan here is imple- have many tunable parameters, and some have more effi- mented with a nodeIDSequence implementation of blocks, cient enterprise versions. For all baseline systems, we map which recall does not perform any data copies. Similarly their storage to an in-memory file system, set each system the join operator is performed using pointers to adjacency to use a single CPU and disable spilling intermediate files lists and finally the group-by operator only counts the sizes to disk. We maintain 2 copies of edge tables for Vertica and of these lists, without any data copies. MonetDB, sorted by the source and destination vertexIDs, respectively. This gives the systems the option to perform 7.7 Baseline System Comparisons fast merge joins (without sorting) when possible instead of Our final set of experiments compares the query perfor- hash joins which require creating hash tables. Because the mance of GF-CL against GF-RV, Neo4j, which is another row- joins in LDBC queries start from a particular vertex ID, the oriented and Volcano style GDBMSs but forms a baseline join order is obvious and we use plans that are left-deep join from an independent code-base, and two column-oriented trees. We noticed that Vertica and MonetDB sometimes did RDBMSs, MonetDB and Vertica, which are not tailored for not pick left-deep plans, in which case we report the better many-to-many joins. Our primary goal is to verify that of the system’s default plan and the left deep plan, which GF-CL is faster than GF-RV also on an independent end- we can force the systems to pick. to-end benchmark. We also aim to verify that the GF-RV Figure 11 shows the relative speedup/slowdown of the dif- system into which we integrated our techniques is already ferent systems in comparison to GF-RV. Table6 shows indi- competitive with or outperforms other baseline systems on vidual runtime numbers of each IS and IC query. As we ex- a workload that contains many-to-many joins. pect, GF-CL is broadly more performant than GF-RV on this We use the LDBC10 dataset. GraphflowDB is a proto- benchmark as well with a median query improvement factor type system and currently implements parts of the Cypher of 2.6x. Importantly, with the exception of one query, which language that are relevant to our research, so lack several slows down slightly, the performance of every query improves features that LDBC queries exercise. The system currently between 1.3x to 8.3x. The improvements come from a combi- has support for select-project-join queries and a limited form nation of optimizations but primarily from LBP and colum- of aggregations, where joins are expressed as fixed-length nar storage of properties. In GF-RV, scanning properties re-

11 IS01 IS02 IS03 IS04 IS05 IS06 IS07 GF-CL 2.7 3.0 2.2 36.9 40.6 69.3 38.3 2.2 3.9 3.9 307.3 236.6 423.0 307.9 GF-RV 0.8x 1.3x 1.8x 8.3 5.8x 6.1x 8.0x 6.1 16728.2 7.0 1.5 45.2 259.2 24818.9 VERTICA 2.2x 5574.2x 3.2x 0.04x 1.1x 3.7x 647.7x 112.3 282.2 8.3 84.1 516.4 323.0 206.3 MONET 40.9x 94.0x 3.8x 2.3x 12.7x 4.7x 5.4x 103.1 117.4 86.1 12418.9 11665.9 67390.3 12095.2 NEO4J 37.5x 39.1x 39.1 336.6 287.4 972.3 315.6 IC01 IC02 IC03 IC04 IC05 IC06 IC07 IC08 IC09 IC011 IC012 GF-CL 36.7 32.4 409.4 13.1 1565.2 113.0 3.0 2.6 1519.8 11.1 34.2 88.4 45.2 1521.8 57.3 8925.0 333.1 6.3 7.0 2098.1 19.2 84.9 GF-RV 2.4x 1.4x 3.7 4.4x 5.7x 3.0x 2.1x 2.7x 1.4x 1.7x 2.5x 257.2 3063.8 18610.3 1711.6 59351.0 4715.7 4092.2 2837.2 17276.2 672.9 5028.1 VERTICA 7.0x 94.5x 45.5x 130.5x 37.9x 41.7x 1348.8x 1094.2x 11.4x 60.9x 147.1x 160.3 323.2 187330.9 13955.1 165273.0 2783.1 206.3 920.5 121943.2 572.0 3251.9 MONET 4.4x 10.0x 457.6x 1064.3x 105.6x 24.6x 68.0x 354.8x 80.0x 51.7x 95.1x 669.3 170722.9 86231.9 75254.9 TLE 515.4 95.6 108.3 219425.5 2804.1 34043.0 NEO4J 18.3x 5264.2x 210.6x 5739.4x - 4.6x 31.5x 41.8x 144.4x 253.6x 996.0x

Table 6: Runtime (in msec) for running the LDBC10 Interactive Complex Reads (IC) and Interactive Short Reads (IS) queries on 5 systems: (i) GF-CL (ii) GF-RV (iii) VERTICA (iv) MONET and (v) NEO4J.

quire performing equality checks on property keys, which and GraphflowDB only scan the Person table to find pi and are avoided in our columnar storages, so we observed that then directly extend pi to its neighbors, avoiding the scan of on queries that perform large intermediate results and per- all Knows edges. In this query, GF-RV, GF-CL, and Neo4j take form filters, such as IC05, we see larger improvements. For 333ms, 113ms, and 515ms, while Vertica and MonetDB take this query, there are 4 many-to-many joins starting from a 4.7s and 2.7s, respectively. We also found that all baseline node and extending in the forward direction and a predicate systems, including Neo4j, degrade in performance on queries on the edges of the third join. GF-CL has several advantages with many many-to-many joins that generate large interme- that become visible here. First, GF-CL’s LBP, unlike GF-RV, diate results. For example, on IC05 that we reviewed above, does not copy any edge and neighbor IDs to intermediate Vertica take 1 minute, MonetDB 3.25 minutes, while Neo4j tuples. More importantly, LBP performs filters inside loops took over 10 minutes. and GF-CL’s single-indexed property pages provides faster access to the edge properties that are used in the filter than 8. RELATED WORK GF-RV’s interpreted attribute layout format. On this query Column stores [41, 58, 63, 64] are designed primarily for GF-RV takes 8.9s while GF-CL takes 1.6s. OLAP applications that perform aggregations over large As we expected, we also found the other baseline systems amounts of data. Work on column stores introduced a set of to not be as performant as GF-CL or GF-RV. In particular Ver- columnar storage, compression, and query processing tech- tica, MonetDB, and Neo4j have median slowdown factors of niques which include use of positional offsets, columnar com- 13.1x, 22.8x, and 46.1x compared to GF-RV. Although Neo4j pression schemes, block-based query processing, late mate- broadly performed slightly worse than other baselines, we rialization, and direct operation on compressed data, among also observed that there were several queries in which it out- others. A detailed survey of these techniques can be found in performed Vertica and MonetDB (but not GF-RV or GF-CL) reference [18]. This paper studies how to integrate some of by a large margin. These were queries that started from a these techniques into in-memory GDBMSs. As we showed, single node as usual, had several many-to-many joins, but some of these techniques, such as traditional block-based did not generate large intermediate results, such as IS02 processors, NULL compression schemes, or direct columnar or IC06. Interestingly on such queries, GDBMSs, so both storage of edge properties require GDBMS-specific adapta- GraphflowDB and Neo4j, have the advantage of using join tions to optimize them for the access patterns of GDBMSs. operators that use the adjacency list indices to extend a Existing GDBMSs and RDF systems adopt a columnar set of partial matches. This is an implementation of index structure often for storing the topologies of graphs. This is nested loop joins that can be highly efficient if the partial done either by using a variant of vanilla adjacency list for- matches that are extended are small in number. For exam- mat or CSR. Instead, systems often use row-oriented struc- ple the first join of IC06 extends a single Person node, say tures to store properties, such as an interpreted attribute pi, to its two-degree friends. In SQL, this is implemented as layout [25]. For example, Neo4j [9] represents the topol- joining a Person table with a Knows table with a predicate ogy of the graph in adjacency lists that are partitioned by on the Person table to select pi. In Vertica or MonetDB this edge labels and stored in linked-lists, where each edge record join is performed using merge or hash joins, which requires points to the next. Similarly, the properties of each vertex the scan of both Person and Knows tables. Instead, Neo4j and edge are stored in a linked-list, where each property

12 record points to the next and encodes the key, data type, efficient as double indexing of adjacency lists in GDBMSs. and value of the property. Similarly, JanusGraph [5] stores RDF-3X [54] is an RDF system that stores a large triple edges in adjacency lists partitioned by edge labels and prop- table that is indexed in 6 B+ tree indexes over each column. erties as consecutive key-value pairs (so in row-oriented for- Similarly, this storage is not as optimized as the native graph mat). JanusGraph uses variable-length encoding when stor- storages found in GDBMSs. ing edges in the adjacency lists. Instead, we use fixed-length Several prior work has introduced novel storage techniques encodings in our compression schemes. DGraph [2] uses a for storing graphs that are optimized for write-heavy, such key-value store to hold adjacency lists as well as properties. as streaming, workloads. These works propose data struc- All of the above native GDBMSs adopt Volcano-style proces- tures that try to achieve the sequential read capabilities of sors. In contrast, our design adopts columnar structures for CSR while being more optimized for writes. Example sys- vertex and edge properties and a block-based processor. In tems and data structures include LiveGraph [62], Aspen [31], addition, we improve on these designs by more compressed LLAMA [49], STINGER [32], and DISTINGER [34]. In this edge and vertex ID representation (these systems use 8 bytes paper we focused on a read-optimized system setting and im- for each ID) and NULL compression. plemented our techniques inside GraphflowDB, which we are There are also several GDBMSs that are developed di- designing as a read-optimized system. We therefore directly rectly on top of an RDBMS or another database system [12], used CSR to store the graph topology but these techniques such as IBM Db2 Graph [59], Oracle Spatial and Graph [12] are complementary to our work. SAP’s graph database [56]. These systems can benefit from Our list groups represent intermediate results in a factor- the columnar techniques in the underlying RDBMS, which ized form. Prior work on query processing in RDBMSs on are however not optimized for graph storage and queries. For factorized representations, specifically FDB [23, 24], repre- example, SAP’s graph engine uses SAP HANA’s columnar- sents intermediate relations as tries, and have operators that storage for edge tables but these tables do not have CSR-like transform tries into other tries. Unlike traditional query pro- structures for storing edges. cessor architectures, processing is not pipelined and all inter- A work that is closely related to ours is GQ-Fast [48]. mediate results are materialized. Instead, operators in LBP GQ-Fast implements a restricted set of SQL called rela- are variants of traditional block-based operators and per- tionship queries that contain joins of tables, similar to path form computations in a pipelined fashion on batches of list- queries followed with aggregations. The system stores rela- s/arrays of data. This paper focuses on integration of colum- tionship tables, i.e., those that represent many-to-many re- nar storage and query processing techniques into GDBMSs. lationships, in CSR-like indices with heavy-weight compres- and we have not studied how to integrate more advanced sion of lists and has a fully pipelined query processor that factorized processing techniques inside GDBMS within the uses query compilation. Therefore, GQ-Fast studies how scope of this paper. This is an interesting direction that is some techniques in GDBMSs, specifically joins using adja- in our immediate future work directions. cency lists can be integrated into an RDBMS. In contrast, we focus on studying how some techniques from columnar 9. CONCLUSIONS RDBMSs can be integrated into GDBMSs. We intended to Column-oriented RDBMSs are read-optimized analytical but could not compare against GQ-Fast because the sys- systems that have introduced several storage and query pro- tem supports a very limited set of queries (e.g., none of cessing techniques to improve the scalability and perfor- the LDBC queries are supported). In addition the publicly mances of RDBMSs. We studied the integration of these available version had several compilation or runtime errors techniques into GDBMSs, which are also read-optimized an- on sample queries and the system is no longer maintained. alytical systems. While some of these techniques can be di- ZipG [45] is a distributed compressed storage engine for rectly applied to GDBMSs, direct adaptation of others can property graphs that can answer queries to retrieve adjacen- be significantly suboptimal in terms of space and perfor- cies as well as vertex and edge properties. ZipG is based on mance. In this paper, we first outlined a set of guidelines and a compressed data structure called Succinct [22]. Succinct desiderata for designing the physical storage layer and query stores semi-structured data that is encoded as a set of key processor of GDBMSs, based on the typical access patterns and list of values. For example, a vertex v’s properties can in GDBMSs which are significantly different than the typi- be stored with the v’s ID as the key and a list of values, cor- cal workloads of columnar RDBMSs. We then presented our responding to each property. Succinct then compresses these design of columnar storage, compression, and query process- files using suffix arrays and several secondary level indices. ing techniques that are optimized for in-memory GDBMSs. Although the authors report good compression rates, unlike Specifically, we introduced a novel list-based query proces- our structures, access to a particular record is not constant sor, which avoids expensive data copies of traditional block- time and requires accessing secondary indexes followed by a based processors and avoids materialization of adjacency binary search, which is slower than our structures. lists in blocks, a new data structure we call single-indexed Several RDF systems also use columnar structures to store property pages and an accompanying edge ID scheme, and a RDF databases. Reference [21] stores data in a set of columns, new application of Jacobson’s bit vector index for compress- where each column store is a set of (subject, object) pairs for ing NULL and empty lists. Through extensive experiments, each unique predicate. However, this storage is not as opti- we demonstrated the scalability and performance benefits of mized as the standard storage in GDBMSs, e.g., the edges our techniques in an actual system implementation. of a particular object are not stored in native CSR or ad- jacency list format. Hexastore [60] improves on the idea of predicate partitioning by defining a column for each RDF el- ement (subject, predicate or object) and sorting the column in 2 possible ways in B+ trees. This is similar but not as

13 10. ACKNOWLEDGEMENTS [28] A. Bonifati, G. Fletcher, H. Voigt, and N. Yakovets. This work was supported in part by an NSERC Discov- Querying Graphs. 2018. ery grant. We thank Lori Paniak for promptly assisting [29] A. Bonifati, P. Furniss, A. Green, R. Harmer, E. Oshurko, and H. Voigt. Schema Validation and Evolution for Graph with many system issues. We also thank Xiyang Feng and Databases. ER, 2019. Guodong Jin for helping with baseline experiments and Sne- [30] P. Bose, M. He, A. Maheshwari, and P. Morin. Succinct hal Mishra for helping with illustrations. orthogonal range search structures on a grid with applications to text indexing. Workshop on Algorithms and 11. REFERENCES Data Structures, 2009. [31] L. Dhulipala, G. E. Blelloch, and J. Shun. Low-latency [1] Amazon Neptune. https://aws.amazon.com/neptune/, graph streaming using compressed purely-functional trees. 2020. SIGPLAN, 2019. [2] DGraph Github Repository. [32] D. Ediger, R. McColl, J. Riedy, and D. A. Bader. https://github.com/dgraph-io/dgraph, 2020. STINGER: High Performance Data Structure for [3] DuckDB. https://duckdb.org/, 2020. Streaming Graphs. HPEC, 2012. [4] Graphflowdb Source Code. [33] O. Erling, A. Averbuch, J. Larriba-Pey, H. Chafi, https://tinyurl.com/pktcna6j, 2020. A. Gubichev, A. Prat, M.-D. Pham, and P. Boncz. The [5] Janus Graph. https://janusgraph.org, 2020. LDBC Social Network Benchmark: Interactive Workload. [6] Memgraph. https://memgraph.com/, 2020. SIGMOD, 2015. [7] MonetDB source code, (Jun2020-SP1). [34] G. Feng, X. Meng, and K. Ammar. DISTINGER: A https://github.com/MonetDB/MonetDB/releases/tag/ Distributed Graph Data Structure for Massive Dynamic Jun2020_SP1_release, 2020. Graph Processing. IEEE Big Data, 2015. [8] Neo4j Blog on Deletions. https://tinyurl.com/4wf9593y, [35] N. Francis, A. Green, P. Guagliardo, L. Libkin, 2020. T. Lindaaker, V. Marsault, S. Plantikow, M. Rydberg, [9] Neo4j Community Edition. P. Selmer, and A. Taylor. Cypher: An Evolving Query https://neo4j.com/download-center/#community, 2020. Language for Property Graphs. SIGMOD, 2018. [10] Neo4j Property Graph Model. [36] J. Goldstein, R. Ramakrishnan, and U. Shaft. Compressing https://neo4j.com/developer/graph-database, 2020. Relations and Indexes. ICDE, 1998. [11] Oracle In-Memory Column Store Architecture. [37] G. Gonnet, R. Baeza-Yates, and T. Snider. New Indices for https://tinyurl.com/vkvb6p6, 2020. Text: Pat Trees and Pat Arrays. 1992. [12] Oracle Spatial and Graph. https://www.oracle.com/ [38] G. Graefe. Volcano: An Extensible and Parallel Query database/technologies/spatialandgraph.html, 2020. Evaluation System. TKDE, 1994. [13] TigerGraph. https://www.tigergraph.com, 2020. [39] G. Graefe and L. Shapiro. Data Compression and Database [14] Vertica 10.0.x Documentation. https://www.vertica.com/ Performance. High-Performance Web Databases, 2001. docs/10.0.x/HTML/Content/Home.html, 2020. [40] O. Hartig and J. Hidders. Defining Schemas for Property [15] GraphflowDB Columnar Techniques. https://github.com/ Graphs by Using the GraphQL Schema Definition g31pranjal/graphflow-columnar-techniques, 2021. Language. GRADES-NDA, 2019. [16] Wikipedia Dynamic (de), (Konect). [41] S. Idreos, F. Groffen, N. Nes, S. Manegold, S. Mullender, http://konect.cc/networks/link-dynamic-dewiki/, 2021. and M. Kersten. MonetDB: Two Decades of Research in [17] D. Abadi. Column Stores for Wide and Sparse Data. Column-Oriented Database Architectures. IEEE Data Eng. CIDR, 2007. Bull., 2012. [18] D. Abadi, P. A. Boncz, S. Harizopoulos, S. Idreos, and [42] G. Jacobson. Space-efficient static trees and graphs. S. Madden. The Design and Implementation of Modern Symposium on Foundations of Computer Science, 1989. Column-Oriented Database Systems. Foundations and [43] G. Jacobson. Succinct static data stttlctates. PhD thesis, Trends in Databases, 2013. Carnegie Mellon University, 1989. [19] D. Abadi, S. Madden, and M. Ferreira. Integrating [44] C. Kankanamge, S. Sahu, A. Mhedbhi, J. Chen, and Compression and Execution in Column-Oriented Database S. Salihoglu. Graphflow: An Active Graph Database. Systems. SIGMOD, 2006. SIGMOD, 2017. [20] D. J. Abadi, S. R. Madden, and N. Hachem. Column-Stores [45] A. Khandelwal, Z. Yang, E. Ye, R. Agarwal, and I. Stoica. Vs. Row-Stores: How Different Are They Really? ZipG: A Memory-Efficient Graph Store for Interactive SIGMOD, 2008. Queries. SIGMOD, 2017. [21] D. J. Abadi, A. Marcus, S. R. Madden, and K. Hollenbach. [46] J. Kunegis. Konect: The koblenz network collection. Scalable Semantic Web Data Management Using Vertical WWW, 2013. Partitioning. PVLDB, 2007. [47] D. Lemire and L. Boytsov. Decoding Billions of Integers Per [22] R. Agarwal, A. Khandelwal, and I. Stoica. Succinct: Second Through Vectorization. Softw. Pract. Exp., 2015. Enabling Queries on Compressed Data. NSDI, 2015. [48] C. Lin, B. Mandel, Y. Papakonstantinou, and M. Springer. [23] N. Bakibayev, T. Kočiský, D. Olteanu, and J. Závodný. Fast in-memory sql analytics on typed graphs. ICDE, 2016. Aggregation and ordering in factorised databases. PVLDB, [49] P. Macko, V. J. Marathe, D. W. Margo, and M. I. Seltzer. 2013. Llama: Efficient graph analytics using large multiversioned [24] N. Bakibayev, D. Olteanu, and J. Závodný. Fdb: A query arrays. ICDE, 2015. engine for factorised relational databases. PVLDB, 2012. [50] A. Mhedhbi and S. Salihoglu. Optimizing Subgraph Queries [25] J. L. Beckmann, A. Halverson, R. Krishnamurthy, and J. F. by Combining Binary and Worst-Case Optimal Joins. Naughton. Extending RDBMSs to Support Sparse Datasets PVLDB, 2019. Using an Interpreted Attribute Storage Format. ICDE, [51] A. Mislove, H. S. Koppula, K. P. Gummadi, P. Druschel, 2006. and B. Bhattacharjee. Growth of the flickr social network. [26] P. Boncz. Monet: A Next-Generation Database Kernel for WOSN, 2008. Query-Intensive Applications. PhD thesis, Universiteit van [52] G. Navarro and V. Mäkinen. Compressed Full-Text Amsterdam, 2002. Indexes. CSUR, 2007. [27] P. Boncz, M. Zukowski, and N. Nes. MonetDB/X100: [53] G. Navarro and L. M. S. Russo. Space-efficient Hyper-Pipelining Query Execution. CIDR, 2005. data-analysis queries on grids. Algorithms and Computation, 2011.

14 [54] T. Neumann and G. Weikum. The RDF3X Engine for 5. IS05 Scalable Management of Rdf Data. VLDBJ, 2010. [55] D. Olteanu and M. Schleich. Factorized Databases. MATCH (comment:Comment)−[:hasCreator]→(p:Person) SIGMOD Rec., 2016. WHERE comment.id = 0 [56] M. Rudolf, M. Paradies, C. Bornhövd, and W. Lehner. The RETURN p.id, p.fName, p.lName; Graph Story of the SAP HANA Database. BTW, 2013. 6. IS06 [57] S. Sahu, A. Mhedhbi, S. Salihoglu, J. Lin, and M. T. Özsu. The Ubiquity of Large Graphs and Surprising Challenges of MATCH (comment:Comment)−[:replyOf]→(pst:Post) Graph Processing: Extended Survey. VLDBJ. (pst:Post)←[:containerOf]−(f:Forum) [58] M. Stonebraker, D. J. Abadi, A. Batkin, X. Chen, (f:Forum)−[:hasModerator]→(p:Person) M. Cherniack, M. Ferreira, E. Lau, A. Lin, S. Madden, WHERE comment.id = 0 E. O’Neil, P. O’Neil, A. Rasin, N. Tran, and S. Zdonik. RETURN f.id, f.title, p.id, p.fName, p.lName; C-store: A Column-oriented DBMS. VLDB, 2005. [59] Y. Tian, E. L. Xu, W. Zhao, M. H. Pirahesh, S. J. Tong, 7. IS07 W. Sun, T. Kolanko, M. S. H. Apu, and H. Peng. Ibm db2 graph: Supporting synergistic and retrofittable graph MATCH (mAuth:Person)←[:hasCreator]−(cmt0:Comment) queries inside ibm db2. 2020. (cmt0:Comment)←[:replyOf]−(cmt1:Comment) [60] C. Weiss, P. Karras, and A. Bernstein. Hexastore: Sextuple (cmt1:Comment)−[:hasCreator]→(rAuth:Person) Indexing for Semantic Web Data Management. PVLDB, WHERE comment.id = 6 2008. RETURN cmt1.id, cmt1.content, cmt1.creationDate, [61] H. Zhang, H. Lim, V. Leis, D. G. Andersen, M. Kaminsky, rAuth.id, rAuth.fName, rAuth.lName; K. Keeton, and A. Pavlo. Succinct range filters. TODS, 2020. 8. IC01 [62] X. Zhu, G. Feng, M. Serafini, X. Ma, J. Yu, L. Xie, MATCH (person:Person)−[:knows]→(p1:Person) A. Aboulnaga, and W. Chen. LiveGraph: A Transactional (p1:Person)−[:knows]→(p2:Person) Graph Storage System with Purely Sequential Adjacency (p2:Person)−[:knows]→(op:Person) List Scans. PVLDB, 2020. (op:Person)−[:isLocatedIn]→(pl:Place) [63] M. Zukowski and P. Boncz. From X100 to Vectorwise: WHERE person.id = 22468883 Opportunities, Challenges and Things Most Researchers Do RETURN op.id, op.lName, op.birthday, op.creationDate, Not Think About. SIGMOD, 2012. op.gender, op.locationIP, city.name; [64] M. Zukowski and P. A. Boncz. Vectorwise: Beyond Column Stores. IEEE Data Eng. Bull., 2012. 9. IC02 [65] M. Zukowski, P. A. Boncz, N. Nes, and S. Héman. MonetDB/X100-A DBMS in the CPU Cache. IEEE Data MATCH (p:Person)−[:knows]→(frnd:Person) Eng. Bull., 2005. (frnd:Person)←[:hasCreator]−(msg:Comment) [66] M. Zukowski, S. Heman, N. Nes, and P. Boncz. WHERE p.id = 22468883 AND Super-Scalar RAM-CPU Cache Compression. ICDE, 2006. msg.creationDate < 1342805711 RETURN frnd.id, frnd.fName, frnd.lName, msg.id, msg.content,msg.creationDate; APPENDIX 10. IC03 A. MODIFIED LDBC SNB QUERIES MATCH (person:Person)−[:knows]→(p1:Person) 1. IS01 (p1:Person)−[:knows]→(op:Person) (op:Person)−[:isLocatedIn]→(pl:Place) MATCH (p:Person)−[:IsLocatedIn]→(pl:Place) (op:Person)←[:hasCreator]−(mx:Comment) WHERE p.id = 22468883 (mx:Comment)−[:isLocatedIn]→(px:Place) RETURN p.fName, p.lName, p.birthday, p.locationIP, (op:Person)←[:hasCreator]−(my:Comment) p.browserUsed, p.gender, p.creationDate, (my:Comment)−[:isLocatedIn]→(py:Place) pl.id; WHERE person.id = 22468883 AND mx.creationDate >= 1313591219 AND 2. IS02 mx.creationDate <= 1513591219 AND my.creationDate >= 1313591219 AND MATCH (p:Person)←[:hasCreator]−(c:Comment) my.creationDate <= 1513591219 AND (c:Comment)−[:replyOf]→(post:Post) cx.name = ’India’ AND cy.name = ’China’ (post:Post)−[:hasCreator]→(op:Person) RETURN cmt1.id, cmt1.content, cmt1.creationDate, WHERE p.id = 22468883 rAuth.id, rAuth.fName, rAuth.lName; RETURN c.id, c.content, c.creationDate, op.id, op.fName, op.lName; 11. IC04

3. IS03 MATCH (:Person)←[:knows]−(p:Person) (p:Person)−[:knows]→(frnd:Person) MATCH (p:Person)−[k:knows]→(friend:Person) (frnd:Person)←[:hasCreator]−(pst:Post) WHERE p.id = 22468883 (pst:Post)−[:hasTag]→(t:Tag) RETURN friend.id, friend.fName, friend.lName, WHERE p.id = 22468883 AND e.date; post.creationDate >= 1313591219 AND 4. IS04 post.creationDate <= 1513591219 RETURN cmt1.id, cmt1.content, cmt1.creationDate, MATCH (comment:Comment) rAuth.id, rAuth.fName, rAuth.lName; WHERE comment.id = 0 RETURN comment.creationDate, comment.content;

15 12. IC05 RETURN cmtAuth.id, cmtAuth.fName, cmtAuth.lName, cmt.creationDate, cmt.id, cmt.content; MATCH (p1:Person)−[:knows]→(p2:Person) (p2:Person)−[:knows]→(p3:Person) 16. IC09 (p3:Person)←[hm:hasMember]−(f:Forum) (f:Forum)−[:containerOf]→(pst:Post) MATCH (p1:Person)−[:knows]→(p2:Person) WHERE p1.id = 22468883 AND hm.date > 1267302820 (p2:Person)−[:knows]→(p3:Person) RETURN f.title; (p3:Person)←[:hasCreator]−(cmt:Comment) WHERE person.id = 22468883 AND 13. IC06 cmt.creationDate < 1342840042 RETURN p3.id, p3.fName, p3.lName, cmt.id, cmt.content, MATCH (p1:Person)−[:knows]→(p2:Person) cmt.creationDate; (p2:Person)−[:knows]→(p3:Person) (p3:Person)←[:hasCreator]−(pst:Post) 17. IC11 (pst:Post)−[:hasTag]→(t1:Tag) (pst:Post)−[:hasTag]→(t2:Tag) MATCH (p1:Person)−[:knows]→(p2:Person) WHERE p1.id = 22468883 AND (p2:Person)−[:knows]→(p3:Person) t1.name = ’Rumi’ AND t2.name <> ’Rumi’ (p3:Person)−[w:workAt]→(org:Organization) RETURN t2.name; (org:Organization)−[:isLocatedIn]→(pl:Place) WHERE p1.id = 22468883 AND w.year < 2016 AND 14. IC07 pl.name = ’China’ RETURN p3.id, p3.fName, p3.lName, org.name; MATCH (p:Person)←[:hasCreator]−(cmt:Comment) (cmt:Comment)←[l:likes]−(frnd:Person) 18. IC12 WHERE p.id = 22468883 RETURN frnd.id, frnd.fName, frnd.lName, l.date, MATCH (p1:Person)−[:knows]→(p2:Person) cmt.content; (p2:Person)←[:hasCreator]−(cmt:Comment) (cmt:Comment)−[:replyOf]−(pst:Post) 15. IC08 (pst:Post)−[:hasTag]→(t:Tag) (t:Tag)−[:hasType]→(tc:TagClass) MATCH (p:Person)←[:hasCreator]−(pst:Post) (tc:TagClass)−[:isSubclassOf]→(:TagClass) (pst:Post)←[:replyOf]−(cmt:Comment) WHERE p1.id = 22468883 AND tc.name=’Person’ (cmt:Comment)−[:hasCreator]→(cmtAuth:Person) RETURN p2.id, p2.fName, p2.lName; WHERE p1.id = 22468883

16