Preliminary concept platform

Smart Objects for Smart Care

SOSC is an early-stage sensing platform designed to transform everyday objects into interactive digital care assets. In this preliminary phase, the project focuses on combining a smartwatch, smart tags, and smart objects to collect rich multimodal data and recognize user–object interactions in daily life.

3-layer system Smartwatch, smart tags, and instrumented objects working together.
Multimodal sensing Motion, physiology, environmental variables, outdoor and indoor context.
Interaction-centric Focused on capturing when, how, and with which object the user interacts.

Smartwatch Hub

Wearable core for motion, physiology, GPS, and continuous personal sensing.

Tagged Objects

Smart tags associated with everyday objects to support user–object interaction recognition.

Cup Fridge Drawer Door

Data Streams

Motion, physiological, environmental, spatial, and interaction data collected in parallel.

Interaction Recognition

Signals from the smartwatch and nearby tagged objects are fused to identify object-related actions.

Project snapshot

What SOSC addresses in the preliminary phase

The current phase is centered on building the technological foundation: defining the sensing ecosystem, collecting synchronized streams from multiple sources, and validating the feasibility of recognizing interactions between users and everyday objects.

01

Smartwatch-centered architecture

The smartwatch acts as the main personal sensing device, continuously collecting data from wrist-worn motion, physiological, and spatial sensors.

02

Smart tags linked to objects

Smart tags are attached to relevant objects in the environment to provide contextual information and support recognition of user–object proximity and interaction.

03

Smart objects as context sources

Objects become digitally traceable entities whose usage can be captured and associated with the user’s wrist signals and spatial patterns.

04

Multimodal data collection for daily-life studies

The platform is designed to gather heterogeneous data streams that can later support activity understanding, context modeling, and more advanced digital care functions.

Layer 1

Smartwatch

Captures continuous personal sensing from the wrist, including movement, orientation, location, and physiology.

Layer 2

Smart Tags

Provide object identity and contextual proximity information for monitored everyday items.

Layer 3

Smart Objects

Enable object-aware monitoring by linking usage events to the user’s behavior and surrounding environment.

Output

Interaction-aware dataset

Generates synchronized multimodal data streams suitable for studying user–object interactions in real settings.

Data foundation

Multimodal signals collected by SOSC

A core goal of the preliminary project is to build a rich and synchronized dataset that combines wearable, environmental, and contextual information from the user and surrounding objects.

Motion sensing

Wrist movement and orientation

The smartwatch captures body movement and wrist dynamics as the user performs object-related actions and daily activities.

  • Wrist acceleration
  • Angular velocity
  • Magnetic field
  • Rotation matrix
Physiological sensing

Personal state indicators

Physiological variables are collected to enrich the understanding of user state during interactions and routines.

  • Heart rate (HR)
  • Heart rate variability (HRV)
Environmental + spatial

Context of the surrounding space

Environmental and positioning information adds context to the interaction events and daily scenarios being monitored.

  • Environment temperature
  • Environment pressure
  • Luminosity
  • GPS
  • Indoor position
Recognition target

User–object interaction recognition

The distinguishing element of SOSC in this preliminary phase is the ability to identify when the user engages with tagged objects by combining wearable data, object identity, proximity, and environmental context.

How interactions are represented

Interactions are modeled through the combination of wrist motion patterns, orientation changes, object proximity, and contextual information from the surrounding environment. This allows the project to distinguish meaningful actions involving specific objects.

Approach Reach Handle Manipulate Release

Why interaction recognition matters

Recognizing how users interact with everyday objects provides a pathway to building richer digital care systems, since object usage often reflects routines, habits, and meaningful components of daily living.

Routine tracking Context awareness Behavior understanding Digital care support
Project phase

Preliminary work package focus

This version of the webpage describes only the initial project phase, where the platform concept, sensing infrastructure, and interaction-aware data collection workflow are being established.

1

Platform setup

Define the sensing ecosystem around the smartwatch, smart tags, and selected smart objects to support synchronized collection in realistic environments.

2

Data acquisition

Collect multimodal streams including motion, physiology, spatial context, and environmental variables during interaction-rich daily scenarios.

3

Interaction feasibility

Assess whether the combined signals enable robust recognition of user–object interactions and provide a strong basis for later project phases.

SOSC starts from the object level

In its preliminary phase, Smart Objects for Smart Care is focused on turning everyday objects into digital signals for care-oriented understanding, starting from sensing, synchronization, and interaction recognition.

Back to top