Proceedings of the 5th Annual on Lifelog Search Challenge 2022
DOI: 10.1145/3512729.3533003
|View full text |Cite
|
Sign up to set email alerts
|

vitrivr at the Lifelog Search Challenge 2022

Abstract: In this paper, we present the iteration of the multimedia retrieval system vitrivr participating at LSC 2022. vitrivr is a general-purpose retrieval system which has previously participated at LSC. We describe the system architecture and functionality, and show initial results based on the test and validation topics.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2
1

Relationship

1
5

Authors

Journals

citations
Cited by 13 publications
(5 citation statements)
references
References 37 publications
0
5
0
Order By: Relevance
“…The most common approach to organizing, retrieving, and analyzing data from wearable cameras involves assigning semantic contexts to images, like visual descriptions, time, and location [ 13 , 14 ]. Various computer vision models are employed to extract visual information from the images, including object detection, activity recognition, optical character recognition [ 13 , 15 ], and embedding models [ 16 , 17 ]. A typical retrieval system would also incorporate different techniques, namely, query enhancement [ 13 ], visual similarity search [ 16 ], and temporal search [ 16 ].…”
Section: Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…The most common approach to organizing, retrieving, and analyzing data from wearable cameras involves assigning semantic contexts to images, like visual descriptions, time, and location [ 13 , 14 ]. Various computer vision models are employed to extract visual information from the images, including object detection, activity recognition, optical character recognition [ 13 , 15 ], and embedding models [ 16 , 17 ]. A typical retrieval system would also incorporate different techniques, namely, query enhancement [ 13 ], visual similarity search [ 16 ], and temporal search [ 16 ].…”
Section: Discussionmentioning
confidence: 99%
“…Various computer vision models are employed to extract visual information from the images, including object detection, activity recognition, optical character recognition [ 13 , 15 ], and embedding models [ 16 , 17 ]. A typical retrieval system would also incorporate different techniques, namely, query enhancement [ 13 ], visual similarity search [ 16 ], and temporal search [ 16 ]. We previously reviewed the SIB image data manually to determine the feasibility and acceptability of wearable cameras to assess self-care in people living with HF [ 21 ].…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…This makes the model more robust against various nuances or typos in the text query and allows users to construct more natural descriptions of the searched image, rather than "keyword-style" descriptions. The impressive performance of CLIP has led to its adoption by teams in the following year [49].…”
Section: B Multimodal Embeddingsmentioning
confidence: 99%
“…Enhancements focused on integrating the CLIP model for retrieval and providing a novice user-friendly search interface. The vitrivr system, a general purpose multimedia data indexing and retrieval framework [8], supported lifelog search using a number of fundamental techniques, such as text embedding, visual similarity search, classical Boolean operators as well as supporting lifelog-specific enhancements such as dynamic result sequencing and faceted filtering. Built on-top of the vitrivr stack was vitrivr-VR [15], which provided a VR-based access mechanism to the lifelog dataset.…”
Section: Participating Systemsmentioning
confidence: 99%