The audio notebook: paper and pen interaction with structured speech
Proceedings of the SIGCHI Conference on Human Factors in Computing Systems
Lifestreams: a storage model for personal data
ACM SIGMOD Record
Proceedings of the 5th international conference on Multimodal interfaces
HLT '01 Proceedings of the first international conference on Human language technology research
Towards event detection in an audio-based sensor network
Proceedings of the third ACM international workshop on Video surveillance & sensor networks
SEVA: sensor-enhanced video annotation
Proceedings of the 13th annual ACM international conference on Multimedia
Unsupervised content discovery in composite audio
Proceedings of the 13th annual ACM international conference on Multimedia
Accessing Minimal-Impact Personal Audio Archives
IEEE MultiMedia
VFerret: content-based similarity search tool for continuous archived video
Proceedings of the 3rd ACM workshop on Continuous archival and retrival of personal experences
Extraction of social context and application to personal multimedia exploration
MULTIMEDIA '06 Proceedings of the 14th annual ACM international conference on Multimedia
Multimodal estimation of user interruptibility for smart mobile telephones
Proceedings of the 8th international conference on Multimodal interfaces
CA3: collaborative annotation of audio in academia
ACM-SE 45 Proceedings of the 45th annual southeast regional conference
Prototyping Applications to Document Human Experiences
IEEE Pervasive Computing
Dynamic privacy assessment in a smart house environment using multimodal sensing
ACM Transactions on Multimedia Computing, Communications, and Applications (TOMCCAP)
SEVA: Sensor-enhanced video annotation
ACM Transactions on Multimedia Computing, Communications, and Applications (TOMCCAP)
Ubiquitous Computing for Capture and Access
Foundations and Trends in Human-Computer Interaction
Unstructured audio classification for environment recognition
AAAI'08 Proceedings of the 23rd national conference on Artificial intelligence - Volume 3
Environmental sound recognition with time-frequency audio features
IEEE Transactions on Audio, Speech, and Language Processing
Text-like segmentation of general audio for content-based retrieval
IEEE Transactions on Multimedia
The CLEAR 2006 CMU acoustic environment classification system
CLEAR'06 Proceedings of the 1st international evaluation conference on Classification of events, activities and relationships
Segmentation, indexing, and retrieval for environmental and natural sounds
IEEE Transactions on Audio, Speech, and Language Processing
Audio-based semantic concept classification for consumer video
IEEE Transactions on Audio, Speech, and Language Processing
Multimodal segmentation of lifelog data
Large Scale Semantic Access to Content (Text, Image, Video, and Sound)
Automatically generating stories from sensor data
Proceedings of the 16th international conference on Intelligent user interfaces
Hi-index | 0.00 |
Collecting and storing continuous personal archives has become cheap and easy, but we are still farfromcreating a useful, ubiquitous memory aid. We view the inconvenience to the user of being 'instrumented'as one of the key barriers to the broader development and adoption of these technologies. Audio-only recordings,however, can have minimal impact, requiring only that a device the size and weight of a cellphone be carried somewhere on the person. We have conducted some small-scale experiments on collecting continuous personal recordings of this kind, and investigating how they can be automatically analyzed and indexed, visualized, and correlated with other minimal-impact, opportunistic data feeds (such as online calendars and digital photo collections). We describe our unsupervised segmentation and clustering experiments in which we can achieve good agreement with hand-marked environment/situation labels. We al so di scuss some of the broader issues raised by this kind of work including privacy concerns,and describe our future plans to address these and other questions.