HLogformer: A Hierarchical Transformer for Representing Log Data
Transformers have gained widespread acclaim for their versatility in handling diverse data structures, yet their application to log data remains underexplored. Log data, characterized by its hierarchical, dictionary-like structure, poses unique challenges when processed using conventional transforme...
Gespeichert in:
Hauptverfasser: | , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Transformers have gained widespread acclaim for their versatility in handling
diverse data structures, yet their application to log data remains
underexplored. Log data, characterized by its hierarchical, dictionary-like
structure, poses unique challenges when processed using conventional
transformer models. Traditional methods often rely on manually crafted
templates for parsing logs, a process that is labor-intensive and lacks
generalizability. Additionally, the linear treatment of log sequences by
standard transformers neglects the rich, nested relationships within log
entries, leading to suboptimal representations and excessive memory usage.
To address these issues, we introduce HLogformer, a novel hierarchical
transformer framework specifically designed for log data. HLogformer leverages
the hierarchical structure of log entries to significantly reduce memory costs
and enhance representation learning. Unlike traditional models that treat log
data as flat sequences, our framework processes log entries in a manner that
respects their inherent hierarchical organization. This approach ensures
comprehensive encoding of both fine-grained details and broader contextual
relationships.
Our contributions are threefold: First, HLogformer is the first framework to
design a dynamic hierarchical transformer tailored for dictionary-like log
data. Second, it dramatically reduces memory costs associated with processing
extensive log sequences. Third, comprehensive experiments demonstrate that
HLogformer more effectively encodes hierarchical contextual information,
proving to be highly effective for downstream tasks such as synthetic anomaly
detection and product recommendation. |
---|---|
DOI: | 10.48550/arxiv.2408.16803 |