Architectural Overview
Architectural Philosophy
Deeptrain is architected as an orchestration layer that sits between fragmented, multi-modal data sources and Large Language Models (LLMs). The core objective of the architecture is to abstract the complexities of data ingestion, processing, and vectorization, providing a unified interface for AI agents to interact with non-textual data.
The system utilizes the VMTP (Video Multi-dimensional Transfer Protocol) to standardize how various data types—ranging from static images to live video streams—are ingested and translated into a format that model-agnostic LLMs can interpret.
The VMTP Data Pipeline
The pipeline follows a structured path from raw data source to model-ready context. This ensures that even models without native vision or audio capabilities can process complex multi-modal inputs.
1. Ingestion Layer (VMTP)
The VMTP protocol acts as the entry point. It handles the handshake between external platforms (YouTube, Vimeo, Local Storage) and the Deeptrain processing engine. It supports both batch processing and real-time streams.
2. Multi-modal Processing Engine
Once data is ingested, it is routed through specialized processing modules:
- Vision Module: Converts flowcharts, graphs, and images into structured descriptive data or intermediate embeddings.
- Audio/Video Module: Utilizes the Transcribe API to convert temporal data (speech and video frames) into sequential text and metadata.
- Text Module: Manages traditional text data, bypassing context window limitations through intelligent chunking.
3. State & Retrieval (Localized Embedding DB)
Processed data is stored in a localized embedding database. This architecture allows the AI agent to perform RAG (Retrieval-Augmented Generation) across multi-modal data types, ensuring the model only receives the most relevant "slices" of data within its context window.
Public Interface & Usage
Users interact with the architecture primarily through the Deeptrain connector and the Transcribe API. The following examples demonstrate how to initialize the pipeline and process multi-dimensional data.
Initializing a Multi-modal Session
To start using Deeptrain, you define your data source and the target model. The system automatically selects the appropriate VMTP driver.
from deeptrain import DeeptrainConnector
# Initialize the connector for a specific model
# Supports 200+ private and open-source models
connector = DeeptrainConnector(model="your-llm-choice")
# Connect a video source via VMTP
connector.add_source(
source_type="video",
url="https://www.youtube.com/watch?v=example",
metadata={"category": "training"}
)
Using the Transcribe API
The Transcribe API is the public interface for handling audio and video inputs. It accepts raw files or URLs and returns processed text and temporal markers.
Input Signature:
source(String/File): Path to local file or URL (YouTube/Vimeo).mode(String): Options includehigh_fidelity,fast, ormulti_dimensional.
# Processing a video input for AI training context
transcription_data = connector.transcribe_api.process(
source="path/to/local/video.mp4",
mode="multi_dimensional"
)
print(transcription_data.text) # Extracted speech and visual descriptions
Contextual Retrieval
Because Deeptrain manages a localized embedding database, users can query their data across modes without manually managing vector stores.
# Querying across flowcharts and video transcripts
query_result = connector.query("Explain the logic in the system flowchart video")
# Output is optimized for your model's context window
response = connector.generate_response(query_result)
Component Specifications
| Component | Role | Public/Internal | | :--- | :--- | :--- | | VMTP Driver | Manages protocol handshakes for external data sources. | Internal | | Transcribe API | Public gateway for video and audio processing. | Public | | Local Embedding DB | Stores vectorized multi-modal content for real-time retrieval. | Internal (Configurable) | | Model Adapter | Normalizes output for 200+ supported LLMs. | Public |
By decoupling the data ingestion (VMTP) from the model inference, Deeptrain allows developers to build "vision-enabled" or "audio-aware" agents using standard text-based LLMs.