Build Status Coverage Status Codacy Rating Maven Central Bintray ScalaDoc Gitter

How Cassandra tables are implemented in Phantom

2.0.0 changed the way the table name is inferred.

Back in the pre 2.0.0 days where reflection was king, the reflection mechanism trivially allowed the final piece of the inheritance chain of a particular table to dictate its name. That means the usual way in which a table name used to get set was by the object that was always part of a Database class.

Let’s consider the below example.


import com.outworkers.phantom.dsl._
import org.joda.time.DateTime

case class Recipe(
  url: String,
  description: Option[String],
  ingredients: List[String],
  servings: Option[Int],
  lastCheckedAt: DateTime,
  props: Map[String, String],
  uid: UUID
)

abstract class Recipes extends Table[Recipes, Recipe] {

  object url extends StringColumn with PartitionKey

  object description extends OptionalStringColumn

  object ingredients extends ListColumn[String]

  object servings extends OptionalIntColumn

  object lastcheckedat extends DateTimeColumn

  object props extends MapColumn[String, String]

  object uid extends UUIDColumn
}

class MyDb(override val connector: CassandraConnection) extends Database[MyDb](connector) {
  object recipes extends Recipes with Connector
}

In the past, when the table was nested within a Database, such as as above, the reflection mechanism would “traverse” the type hierarchy and work out that the last thing in the chain was the recipes object, and as a result it would use that as the name of the table within Cassandra.

However, the macro engine will instead try to figure out the last type in the type hierarchy that directly defines columns.

This enables inheritance, but it does not support singletons/objects, so as a result, in an identical scenario, the macro engine will infer the table name as “Recipes”, based on the type information. If you hit trouble upgrading because names no longer match, simply override the table name manually inside the table definition.


import com.outworkers.phantom.dsl._

class MyDb(override val connector: CassandraConnection) extends Database[MyDb](connector) {
  object recipes extends Recipes with Connector {
    override def tableName: String = "recipes"
  }
}

The name of the table can be controlled using NamingStrategy.

A single import controls how table names are generated. Phantom offers three variants, implemented via com.outworkers.phantom.NamingStrategy. These control only the tableName, not the columns or anything else.

Strategy Casing
NamingStrategy.CamelCase lowCamelCase
NamingStrategy.SnakeCase low_snake_case
NamingStrategy.Default Preserves the user input

All available imports will have two flavours. It’s important to note they only work when imported in the scope where tables are defined. That’s where the macro will evaluate the call site for implicits.

import com.outworkers.phantom.NamingStrategy.CamelCase.caseSensitive
import com.outworkers.phantom.NamingStrategy.CamelCase.caseInsensitive

import com.outworkers.phantom.NamingStrategy.SnakeCase.caseSensitive
import com.outworkers.phantom.NamingStrategy.SnakeCase.caseInsensitive

import com.outworkers.phantom.NamingStrategy.Default.caseSensitive
import com.outworkers.phantom.NamingStrategy.Default.caseInsensitive

Data modeling with phantom

back to top


import java.util.UUID
import com.outworkers.phantom.dsl._

case class ExampleModel (
  id: UUID,
  name: String,
  props: Map[String, String],
  timestamp: Int,
  test: Option[Int]
)

abstract class ExampleRecord extends Table[ExampleRecord, ExampleModel] {
  object id extends UUIDColumn with PartitionKey
  object timestamp extends DateTimeColumn with ClusteringOrder with Ascending
  object name extends StringColumn
  object props extends MapColumn[String, String]
  object test extends OptionalIntColumn
}

Extractors and how they get generated

back to top

As of phantom 2.0.0, phantom attempts to automatically derive an appropriate extractor for your record type. In plain English, let’s take the above example with ExampleModel and ExampleRecord. In the olden days, there was no automation, so you would’ve had to manually tell phantom how to produce an ExampleModel from com.datastax.driver.core.Row.

It would look kind of like this:

  def fromRow(row: Row): ExampleModel = {
    new ExampleModel(
      id(row),
      name(row),
      props(row),
      timestamp(row),
      test(row)
    )
  }

Which is just boilerplate, because you already have the schema defined using the modelling DSL, so you don’t really want another layer of manual work.

Extractor derivation and limitations

In 2.0.0, we have addressed this using macros, specifically a macro called com.outworkers.phantom.macros.TableHelper. This class is automatically “computed” using the macro framework, at compile time, and invisibly summoned using implicit macros. You don’t really need to know the inner workings here all the time, as it’s designed to be an invsible background part of the toolkit, but sometimes you may run into trouble, as it’s not perfect.

How the macro works

The macro is capable of matching columns based on their type, regardless of which order they are written in. So even if your record types match the table column types in an arbitrary order, the macro can figure that out. It’s just like in the above example, where ExampleModel doesn’t follow the same order as ExampleRecord.

So the algorithm for the match is very trivial and it works like this:

Field arity
Limitations

The macro will attempt to match multiple fields of the same type based on their name. So if you have two uuids in both your table and your record, they are matched perfectly if their names are identical. At this point in time, phantom 2.4.0, no attempt is made to match fields using string distance algorithms, but we do anticipate this coming to a future version. At this point in time, if a direct match for the field name is not found, then the “next” field in the order written by the user is selected. Most of the time, this is the desired behaviour, but it can fail at runtime because it will mix up your columns

Automatically derived “store” method

back to top

As of phantom 2.5.0, phantom will automatically infer the appropiate store method definition for a combination of Table and Record type arguments. This means you no longer have to manually do anything at all and you can just insert records ad-hoc using pre-generated things.

The inner workings of this generation are non trivial, but the TLDR is quite simple. Know how you extends CassandraTable[TableType, RecordType]?`

Let’s analyse the most common examples and see how this would work.

The simple case

This is the most standard use case, where your table has the exact same number of columns as your record and there is a perfect mapping(bijection) between your table and your record. In this case, the generated store method will simply take a single argument of type Record, as illustrated below.


import com.outworkers.phantom.dsl._
import scala.concurrent.duration._
import com.outworkers.phantom.builder.query.InsertQuery

case class MyRecord(
  id: java.util.UUID,
  name: String,
  firstName: String,
  email: String
)

abstract class MyTable extends Table[MyTable, MyRecord] {

  object id extends UUIDColumn with PartitionKey
  object name extends StringColumn
  object firstName extends StringColumn
  object email extends StringColumn

  // Phantom now auto-generates the below method
  def store(record: MyRecord): InsertQuery.Default[MyTable, MyRecord] = {
    insert.value(_.id, record.id)
      .value(_.name, record.name)
      .value(_.firstName, record.firstName)
      .value(_.email, record.email)
  }

}

The complicated use case

Sometimes we need to store a given Record in more than one table to achieve denormalisation. This is fairly trivial and standard in Cassandra, but it introduces a subtle problem, namely that the new table needs to store a Record and at least one more column representing an index.

We refer to the columns that exist in the Table but not in the Record as unmatched columns, both in the documentation as well as the schema.

Let’s picture we are trying to store Record grouped by countryCode, because we want to be able to query records belonging to a particular countryCode. The macro will pick up on the fact that our table now has more columns than our Record type needs, which means they need to somehow be mapped.

So the new type of the generated store method will now be:

  def store(
    countryCode: String,
    record: Record
  ): InsertQuery.Default[RecordsByCountry, Record]   

This is better visible below, where both the body of the new store method as well as the Cassandra table DSL schema we might use for it are clearly visible.

Warning!!, the order in which the columns are written inside the table is irrelevant, by design the macro will take all that columns that exist in the table but not in the Record and put them in front of the Record type inside the store method input type signature.

The macro will always create a Tuple as described initially, of all the types of unmatched columns, succeeded by the Record type.


import java.util.UUID
import com.outworkers.phantom.dsl._
import scala.concurrent.duration._
import com.outworkers.phantom.builder.query.InsertQuery

case class CountryRecord(
  id: java.util.UUID,
  name: String,
  firstName: String,
  email: String
)

abstract class RecordsByCountry extends Table[RecordsByCountry, CountryRecord] {
  object countryCode extends StringColumn with PartitionKey
  object id extends UUIDColumn with PrimaryKey
  object name extends StringColumn
  object firstName extends StringColumn
  object email extends StringColumn

  // Phantom now auto-generates the below method
  def storeExample(
    countryCode: String,
    record: CountryRecord
  ): InsertQuery.Default[RecordsByCountry, CountryRecord] = {
    insert
      .value(_.countryCode, countryCode)
      .value(_.id, record.id)
      .value(_.name, record.name)
      .value(_.firstName, record.firstName)
      .value(_.email, record.email)
  }

}

To see how this logic might be further extended, let’s add a region partition key to create a Compound primary key that would allow us to retrieve all records by both country and region.

So the new type of the generated store method will now be:

  def store(
    countryCode: String,
    region: String,
    record: Record
  ): InsertQuery.Default[RecordsByCountry, Record]   

The new table definition to store the above is:


import com.outworkers.phantom.dsl._
import com.outworkers.phantom.builder.query.InsertQuery
import scala.concurrent.duration._

case class Record(
  id: UUID,
  name: String,
  firstName: String,
  email: String
)

abstract class RecordsByCountryAndRegion extends Table[RecordsByCountryAndRegion, Record] {
  object countryCode extends StringColumn with PartitionKey
  object region extends StringColumn with PartitionKey
  object id extends UUIDColumn with PrimaryKey
  object name extends StringColumn
  object firstName extends StringColumn
  object email extends StringColumn

  // Phantom now auto-generates the below method
  def storeExample(
    countryCode: String,
    region: String,
    record: Record
): InsertQuery.Default[RecordsByCountryAndRegion, Record] = {
    
    insert
      .value(_.countryCode, countryCode)
      .value(_.region, region)
      .value(_.id, record.id)
      .value(_.name, record.name)
      .value(_.firstName, record.firstName)
      .value(_.email, record.email)
  }

}