Reading view

There are new articles available, click to refresh the page.

WatchThis: A Wearable Point-and-Ask Interface Powered by Vision-Language Models and XIAO ESP32S3 Sense

MIT Media Lab researchers Cathy Mengying Fang, Patrick Chwalek, Quincy Kuang, and Pattie Maes have developed WatchThis, a groundbreaking wearable device that enables natural language interactions with real-world objects through simple pointing gestures. Cathy conceived the idea for WatchThis during a one-day hackathon in Shenzhen, organized as part of MIT Media Lab’s “Research at Scale” initiative. Organized by Cedric Honnet and hosted by Southern University of Science and Technology and Seeed Studio, the hackathon provided the perfect setting to prototype this innovative device using components from the Seeed Studio XIAO ESP32S3 suite. By integrating Vision-Language Models (VLMs) with a compact wrist-worn device, WatchThis allows users to ask questions about their surroundings in real-time, making contextual queries as intuitive as pointing and asking.

Credit: Cathy Fang

Hardwares

The WatchThis project utilizes the following hardware components:

Credit: Cathy Fang

How the Project Works

WatchThis is designed to seamlessly integrate natural, gesture-based interaction into daily life. The wearable device consists of a watch with a rotating, flip-up camera attached to the back of a display. When the user points at an object of interest, the camera captures the area, and the device processes contextual queries based on the user’s gesture.

The interaction begins when the user flips up the watch body to reveal the camera, which then captures the area where the finger points at. The watch’s display shows a live feed from the camera, allowing precise aiming. When the user touches the screen, the device captures the image and pauses the camera feed. The captured RGB image is then compressed into JPG format and converted to base64, after which an API request is made to query the image.

The device uses these API calls to interact with OpenAI’s GPT-4o model, which accepts both text and image inputs. This allows the user to ask questions such as “What is this?” or “Translate this,” and receive immediate responses. The text response is displayed on the screen, overlaid on the captured image. After the response is shown for 3 seconds, the screen returns to streaming the camera feed, ready for the next command.

The software driving WatchThis is written in Arduino-compatible C++ and runs directly on the device. It is optimized for quick and efficient performance, with an end-to-end response time of around 3 seconds. Instead of relying on voice recognition or text-to-speech—which can be error-prone and resource-intensive—the system uses direct text input for queries. Users can further personalize their interactions by modifying the default query prompt through an accompanying WebApp served on the device, allowing tailored actions such as identifying objects, translating text, or requesting instructions.

Credit: Cathy Fang

Applications

Imagine strolling through a city and pointing at a building to learn its history, or identifying an exotic plant in a botanical garden with a mere gesture.

The device goes beyond simple identification, offering practical applications like real-time translation of, for example, menu items, which is a game-changer for travelers and language learners alike.

The research team has discussed even more exciting potential applications:

    • A “Remember this” function could serve as a visual reminder system, potentially aiding those who need to take medication regularly.
    • For urban explorers, a “How do I get there” feature could provide intuitive, spatially-aware navigation by allowing users to point at distant landmarks.
    • A “Zoom in on that” capability could offer a closer look at far-off objects without disrupting the user’s activities.
    • Perhaps most intriguingly, a “Turn that off” function could allow users to control smart home devices with a combination of voice commands and gestures, seamlessly integrating with IoT ecosystems.

While some of these features are still in conceptual stages, they paint a picture of a future where our interactions with the world around us are more intuitive, informative, and effortless than ever before.

Credit: Cathy Fang

Build Your Own WatchThis

Interested in building your own WatchThis wearable? Explore the open-source hardware and software components on GitHub and start creating today! Check out their paper below for full details.

End Note

Hey community, we’re curating a monthly newsletter centering around the beloved Seeed Studio XIAO. If you want to stay up-to-date with:

🤖 Cool Projects from the Community to get inspiration and tutorials
📰 Product Updates: firmware update, new product spoiler
📖 Wiki Updates: new wikis + wiki contribution
📣 News: events, contests, and other community stuff

Please click the image below👇 to subscribe now!

The post WatchThis: A Wearable Point-and-Ask Interface Powered by Vision-Language Models and XIAO ESP32S3 Sense appeared first on Latest Open Tech From Seeed.

AIoT Xperience Day in Düsseldorf: Unlock the Power of AI and IoT

Hey community! We’re excited to invite you to join us at the AIoT Xperience Day in Düsseldorf, an event co-organized by Techbros, Uplink, Heliotics, Infinetics and Seeed Studio

This is an event you should not miss. Either join the hands-on workshops, play with live demos, show your projects or talk with like-minded enthusiasts, innovators, and professionals in the fields of AI and IoT, it is ensured to be a half-day full of excitement!

Hands-on Workshops

During the event, you’ll get the opportunity to explore two exciting workshops:

1. Meshtastic Workshop

Learn how to set up and utilize Meshtastic, a decentralized LoRa mesh network, with the SenseCAP T1000-E Tracker. Whether you’re a beginner or an experienced developer, this session will demonstrate how you can build a free, off-grid communication network to transmit messages without relying on traditional infrastructure.

2. No-Code AI with XIAO ESP32S3 Sense

Can AI be added to almost anything? Yes, and without any coding! In this hands-on workshop, you’ll discover how to easily build your own AI Vision sensor using XIAO ESP32S3 Sense and SenseCraft AI, our no-code platform for training and deploying AI models. With just a few steps, you’ll have your own custom AI Vision sensor up and running. For this workshop participants need to bring their own laptop.

Why Attend?

This event is more than just workshops. You’ll also get to play with live demos of cutting-edge AIoT technology and connect with others who are passionate about the future of AI and IoT.  Don’t miss the “show and tell” session, where participants will have the chance to share their own innovative projects. And you are invited to bring your projects to the spotlight as well!

Whether you’re interested in creating decentralized communication networks, or you want to see how AI can be seamlessly integrated into your projects without a single line of code, this is the event for you.

Event Details

Time: September 27, 2024  2:00 PM – 6:00PM

Venue: Techbros Office, Heerdter Lohweg 212, 40549 Düsseldorf, Germany

Food and drinks will be prepared according to sign-ups, and the spots are limited. Do sign up now to secure your spot! See you in Düsseldorf!

The post AIoT Xperience Day in Düsseldorf: Unlock the Power of AI and IoT appeared first on Latest Open Tech From Seeed.

❌