Schema-on-Read

What is Schema-on-Read?

Schema-on-Read is a data processing strategy where the schema is applied only when reading the data. In contrast to traditional, schema-on-write approaches where data is pre-processed and pre-aggregated before storing, Schema-on-Read allows raw data to be dumped into the data storage system, such as a data lake, without any transformations. The schema, or structure, is then applied when the data is read for analysis.

Functionality and Features

Schema-on-Read offers significant flexibility and adaptability, especially in big data environments. Key features include

  • Flexibility of defining the schema at the time of reading, allowing for diverse data exploration.
  • Ability to handle diverse data types, including structured, semi-structured, and unstructured data.
  • Storage of raw data, enabling full preservation of data, and comprehensive analytics.

Benefits and Use Cases

The Schema-on-Read approach provides various benefits including:

  • Inexpensive storage: As raw data requires less preparation before storage, it reduces data storage costs.
  • Data freshness: Since data isn’t transformed before writing, it can be retrieved in its most original state, ensuring data freshness.
  • Agility: It allows for agile data analysis because the schema can be easily changed as per analysis requirements.

Challenges and Limitations

Despite its benefits, there are some challenges and limitations associated with Schema-on-Read:

  • Performance trade-offs: As data processing occurs at the time of reading data, it might cause delays particularly with large volumes of data.
  • Requires advanced skills: To devise a schema at the time of reading requires advanced technical skills.

Integration with Data Lakehouse

In a data lakehouse, Schema-on-Read plays an integral role by enabling the storage of vast, raw data that can be later processed and read as needed. This greatly enhances the flexibility and efficiency of data analytics, making it a favorable approach in such environments.

Security Aspects

While Schema-on-Read itself doesn't contain specific security measures, the data lake or database systems where it is implemented must ensure data security through access controls, encryption, and other security practices.

Performance

Schema-on-Read can impact performance negatively in some scenarios due to the late binding of schema. However, with the efficient use of computational resources and advanced analytical tools, these potential delays can be mitigated.

FAQs

What is Schema-on-Read? It's a data processing strategy that involves applying the schema at the time of reading data.

What are the benefits of Schema-on-Read? Benefits include inexpensive storage, data freshness, and agility in data analysis.

What are the challenges of Schema-on-Read? Challenges involve performance trade-offs and the requirement of advanced technical skills.

How does Schema-on-Read integrate with data lakehouse? It enables the storage of vast, raw data in a data lakehouse, which can be processed and read as needed.

Does Schema-on-Read impact performance? Yes, it can, particularly with large volumes of data, but this can be mitigated with efficient resource use and advanced tools.

Glossary

Data Lake: A large storage repository that holds raw data in its native format until needed.

Schema-on-Write: A data processing approach where schema is applied at the time of writing data.

Data Lakehouse: A combination of data lake and data warehouse, offering the benefits of both.

Data Freshness: A concept that indicates how recent the data is.

Agile Data Analysis: A flexible approach to data analysis that allows for rapid changes based on evolving requirements.

get started

Get Started Free

No time limit - totally free - just the way you like it.

Sign Up Now
demo on demand

See Dremio in Action

Not ready to get started today? See the platform in action.

Watch Demo
talk expert

Talk to an Expert

Not sure where to start? Get your questions answered fast.

Contact Us

Ready to Get Started?

Enable the business to create and consume data products powered by Apache Iceberg, accelerating AI and analytics initiatives and dramatically reducing costs.