Tecton Adds Low-Latency Streaming Pipelines to its Feature Store

Tecton, the enterprise feature store company, is adding low-latency streaming pipelines to its feature store so that organizations can quickly and reliably build real-time machine learning models.

“Enterprises are increasingly deploying real-time ML to support new customer-facing applications and to automate business processes,” said Kevin Stumpf, co-founder and CTO of Tecton. “The addition of low-latency streaming pipelines to the Tecton feature store enables our customers to build real-time ML applications faster, and with more accurate predictions.”

Real-time ML means that predictions are generated online, at low latency, using an organization’s real-time data; any updates in the data sources are reflected in real-time in the model’s predictions, according to the vendor. Real-time ML is valuable for any use case that is sensitive to the freshness of the predictions, such as fraud detection, product recommendations and pricing use cases.

With Tecton, data teams can build and deploy features using streaming data sources like Kafka or Kinesis in hours, according to the vendor. Users only need to provide the data transformation logic using powerful Tecton primitives, and Tecton executes this logic in fully-managed operational data pipelines which can process and serve features in real-time.

Tecton also processes historical data to create training datasets and backfills that are consistent with the online data and eliminates training/serving skew. Time window aggregations—by far the most common feature type used in real-time ML applications—are supported out-of-the-box with an optimized implementation.

Data teams who are already using real-time ML can now build and deploy models faster, increase prediction accuracy and reduce the load on engineering teams.

Data teams that are new to streaming can build a new class of real-time ML applications that require ultra-fresh feature values. Tecton simplifies the most difficult step in the transition to real-time ML - building and operating the streaming ML pipelines.

For more information about this news, visit