Writing a Single Topic¶
This guide demonstrates how to ingest data into the Mosaico Data Platform from custom files. Here the example of a CSV file is provided, but the logic is compatible with any file format and I/O library. You will learn how to use the Mosaico SDK for:
- Opening a connection to the Mosaico server.
- Creating a sequence.
- Creating a topic.
- Pushing data into a topic.
Step 1: Chunked Loading for High-Volume Data¶
In this example, we assume our CSV file contains the following columns:
timestamp, acc_x, acc_y, acc_z, gyro_x, gyro_y, gyro_z
1110022, 0.0032, 0.001, -0.002, 0.01, 0.005, -0.003
1111022, 0.0041, 0.002, -0.001, 0.012, 0.006, -0.004
1112022, 0.0028, 0.0005, -0.003, 0.009, 0.004, -0.002
The implementation below uses pandas to stream the data, but the logic is compatible with any streaming I/O library.
When dealing with massive datasets, we adopt a chunked loading approach for each sensor type.
import pandas as pd
from mosaicolabs import (
MosaicoClient, # The gateway to the Mosaico Platform
OnErrorPolicy, # The error policy for the SequenceWriter
Message, # The base class for all data messages
IMU, # The IMU sensor data class
Vector3d, # The 3D vector class, needed to populate the IMU data
)
def stream_imu_from_csv(file_path: str, chunk_size: int = 1000):
for chunk in pd.read_csv(file_path, chunksize=chunk_size): # (1)!
for row in chunk.itertuples(index=False):
try:
yield Message(
timestamp_ns=int(row.timestamp),
data=IMU(
acceleration=Vector3d(
x=float(row.acc_x),
y=float(row.acc_y),
z=float(row.acc_z),
),
angular_velocity=Vector3d(
x=float(row.gyro_x),
y=float(row.gyro_y),
z=float(row.gyro_z),
),
),
)
except Exception:
# Yield None only for parsing/type-related errors
yield None
- Use pandas TextFileReader to stream the file in chunks
The Mosaico Message object is an in-memory object wrapping the sensor data with necessary metadata (e.g. timestamp), and ensuring it is ready for serialization and network transmission.
In this specific case, the data is an instance of the IMU model. This is a built-in part of the Mosaico default ontology, meaning the platform already understands its schema and how to optimize its storage.
For a more in-depth explanation:
Step 2: Orchestrating the Sequence Upload¶
To write data, we first establish a connection to the Mosaico server via the MosaicoClient.connect() method and create a SequenceWriter.
A sequence writer acts as a logical container for related data streams (topics).
When initializing your data handling pipeline, it is highly recommended to wrap the MosaicoClient within a with statement. This context manager pattern ensures that underlying network connections and shared resource pools are correctly shut down and released when your operations conclude.
with MosaicoClient.connect("localhost", 6726) as client:
# Initialize the Sequence Orchestrator
with client.sequence_create(
sequence_name="csv_ingestion_test",
metadata={"source": "manual_upload", "format": "csv"}
on_error = OnErrorPolicy.Delete # (1)!
) as swriter:
# Step 3 and 4 happen inside this block...
- Mosaico supports two distinct error policies for sequences:
OnErrorPolicy.DeleteandOnErrorPolicy.Report.
Context Management
It is mandatory to use the SequenceWriter instance returned by client.sequence_create() inside its own with context. The following code will raise an exception:
Sequence-Level Error Handling¶
The behavior of the orchestrator during a failure is governed by the on_error policy. This is a Last-Resort automated error policy, which dictates how the server manages a sequence if an unhandled exception bubbles up to the SequenceWriter context manager. By default, this is set to OnErrorPolicy.Delete, which signals the server to physically remove the incomplete sequence and its associated topic directories, if any errors occurred. Alternatively, you can specify OnErrorPolicy.Report: in this case, the SDK will not delete the data but will instead send an error notification to the server, allowing the platform to flag the sequence as failed while retaining whatever records were successfully transmitted before the error occurred.
For a more in-depth explanation:
Step 3: Topic Creation¶
Inside the sequence, we create a Topic Writer, which is assigned to the IMU topic.
with client.sequence_create(...)
imu_twriter = swriter.topic_create( # (1)!
topic_name="sensors/imu",
metadata={"sensor_id": "accel_01"},
ontology_type=IMU,
)
- Here we are creating a dedicated writer for the IMU topic
Step 4: Pushing Data into the Pipeline¶
The final stage of the ingestion process involves iterating through your data generators and transmitting records to the Mosaico platform by calling the TopicWriter.push() method for each record. The push() method optimizes the throughput by accumulating messages into internal batches.
with client.sequence_create(...)
imu_twriter = swriter.topic_create(...)
for msg in stream_imu_from_csv("imu_data.csv"):
if msg is None:
# Log and skip, or raise if incomplete data is disallowed
print("Skipping row due to parsing error")
continue # Ignore malformed records
try:
imu_twriter.push(message=msg)
except Exception as e:
# Log and skip, or raise if incomplete data is disallowed
print(f"Error at time: {msg.timestamp_ns}. Inner err: {e}")
Topic-Level Error Management¶
In the code snippet above, we implemented a Controlled Ingestion by wrapping the topic-specific processing and pushing logic within a local try-except block.
Because the SequenceWriter cannot natively distinguish which specific topic failed within your custom processing code (such as a coordinate transformation or a malformed CSV row), an unhandled exception will bubble up and trigger the global sequence-level error policy. To avoid this, you should catch errors locally for each topic.
Upcoming versions of the SDK will introduce native Topic-Level Error Policies. This feature will allow you to define the error behavior directly when creating the topic, removing the need for boilerplate try-except blocks around every sensor stream.
The full example code¶
"""
Import the necessary classes from the Mosaico SDK.
"""
import pandas as pd
from mosaicolabs import (
MosaicoClient, # The gateway to the Mosaico Platform
OnErrorPolicy, # The error policy for the SequenceWriter
Message, # The base class for all data messages
IMU, # The IMU sensor data class
Vector3d, # The 3D vector class, needed to populate the IMU data
)
"""
Define the generator functions that yield `Message` objects.
"""
def stream_imu_from_csv(file_path: str, chunk_size: int = 1000):
"""
Efficiently reads a large CSV in chunks to prevent memory exhaustion.
"""
# Use pandas TextFileReader to stream the file in chunks
for chunk in pd.read_csv(file_path, chunksize=chunk_size):
for row in chunk.itertuples(index=False):
try:
yield Message(
timestamp_ns=int(row.timestamp),
data=IMU(
acceleration=Vector3d(
x=float(row.acc_x),
y=float(row.acc_y),
z=float(row.acc_z),
),
angular_velocity=Vector3d(
x=float(row.gyro_x),
y=float(row.gyro_y),
z=float(row.gyro_z),
),
),
)
except Exception:
# Yield None only for parsing/type-related errors
yield None
"""
Main ingestion orchestration
"""
def main():
with MosaicoClient.connect("localhost", 6726) as client:
# Initialize the Sequence Orchestrator
with client.sequence_create(
sequence_name="csv_ingestion_test",
metadata={"source": "manual_upload", "format": "csv"}
on_error = OnErrorPolicy.Delete # Default
) as swriter:
# Create a dedicated writer for the IMU topic
imu_twriter = swriter.topic_create(
topic_name="sensors/imu",
metadata={"sensor_id": "accel_01"},
ontology_type=IMU,
)
# --- Push IMU Data ---
for msg in stream_imu_from_csv("imu.csv"):
if msg is None:
# Log and skip, or raise if incomplete data is disallowed
print("Skipping row due to parsing error")
continue # Ignore malformed records
try:
imu_twriter.push(message=msg)
except Exception as e:
# Log and skip, or raise if incomplete data is disallowed
print(f"Error processing IMU at time: {msg.timestamp_ns}. Inner err: {e}")
# All buffers are flushed and the sequence is committed when exiting the SequenceWriter 'with' block
print("Successfully injected data from CSV into Mosaico!")
# Here the `MosaicoClient` context and all connections are closed