Local Outlier Factor for Feature‐evolving Data Streams: Unterschied zwischen den Versionen
(Die Seite wurde neu angelegt: „{{Vortrag |vortragender=Elena Astankow |email=elena.astankow@student.kit.edu |vortragstyp=Proposal |betreuer=Florian Kalinke |termin=Institutsseminar/2021-01-0…“) |
Keine Bearbeitungszusammenfassung |
||
Zeile 1: | Zeile 1: | ||
{{Vortrag | {{Vortrag | ||
|vortragender=Elena | |vortragender=Elena Schediwie | ||
|email=elena.astankow@student.kit.edu | |email=elena.astankow@student.kit.edu | ||
|vortragstyp=Proposal | |vortragstyp=Proposal | ||
|betreuer=Florian Kalinke | |betreuer=Florian Kalinke | ||
|termin=Institutsseminar/2021-01-08 | |termin=Institutsseminar/2021-01-08 | ||
|kurzfassung= | |kurzfassung=In high-volume data streams it is often unpractical to monitor all observations when we are interested in deviation from normal operations. Detecting outlying observations in data streams is an active area of research. | ||
However, most approaches assume that the data's dimensionality, i.e., the number of attributes, stays constant over time. This assumption is unjustified in many real-world use cases, such as sensor networks or computer cluster monitoring. | |||
Feature-evolving data streams do not impose this restriction and thereby pose additional challenges. | |||
In this thesis, we extend the well-known Local Outlier Factor (LOF) algorithm for outlier detection from the static case to the feature-evolving setting. Our algorithm combines subspace projection techniques with an appropriate index structure using only bounded computational resources. By discarding old observations our approach also deals with concept drift. | |||
We evaluate our approach against the respective state-of-the-art methods in the static case, the streaming case, and the feature-evolving case. | |||
}} | }} |
Version vom 4. Januar 2021, 01:18 Uhr
Vortragende(r) | Elena Schediwie | |
---|---|---|
Vortragstyp | Proposal | |
Betreuer(in) | Florian Kalinke | |
Termin | Fr 8. Januar 2021 | |
Vortragssprache | ||
Vortragsmodus | ||
Kurzfassung | In high-volume data streams it is often unpractical to monitor all observations when we are interested in deviation from normal operations. Detecting outlying observations in data streams is an active area of research.
However, most approaches assume that the data's dimensionality, i.e., the number of attributes, stays constant over time. This assumption is unjustified in many real-world use cases, such as sensor networks or computer cluster monitoring. Feature-evolving data streams do not impose this restriction and thereby pose additional challenges. In this thesis, we extend the well-known Local Outlier Factor (LOF) algorithm for outlier detection from the static case to the feature-evolving setting. Our algorithm combines subspace projection techniques with an appropriate index structure using only bounded computational resources. By discarding old observations our approach also deals with concept drift. We evaluate our approach against the respective state-of-the-art methods in the static case, the streaming case, and the feature-evolving case. |