日本語

Building a Type-safe Data Pipeline

A data pipeline consolidates different data sources, transforms and normalizes it, and then exposes it in a useful way to the customer. With data arriving in all shapes and sizes, how do we maintain reliability and build trust for customers of the pipeline? At Twilio, we enforce schemas for all resources that enter the pipeline, letting us guarantee integrity of the data as defined by the customer’s schema. In this talk, we give an overview of how our Data Platform team has used Scala technologies like Kafka and Spark to build a scalable pipeline, and the benefits of adhering to schemas.

Session length
15 minutes
Language of the presentation
English
Target audience
Intermediate: Requires a basic knowledge of the area
Speaker
Connie Chen (Twilio)

voted / votable

Candidate sessions