Control YouTube 360 videos with head movements and voice commands.
360 VisionVox enhances the experience of watching YouTube 360-degree videos by enabling control through AI-powered voice commands and head movements. It utilizes TensorFlow.js for pose estimation and speech recognition, providing seamless interaction with the video player.
- Voice Commands: Control video playback with voice commands such as Play, Pause, Stop, Next, and Back.
- Head Movement Controls: Navigate 360-degree videos by moving your head in different directions (Up, Down, Left, Right).
- TensorFlow.js: Utilized for both pose estimation (head movement tracking) and speech recognition (audio classifier model).
- GitHub Actions: Used for automated workflows, including deployment tasks.
- HTML5/CSS3/JavaScript: Front-end development technologies for building the user interface and interaction logic.
-
Clone the repository:
git clone https://github.com/lucianoayres/360-visionvox.git
-
Run the application:
Open
index.html
in a web browser or set up a local server.
- Open
index.html
in a web browser. - Ensure your browser has access to the microphone and camera.
- Say (Play, Pause, Stop, Next, Back) to control video playback.
- Enable head movements to navigate 360-degree videos:
- Look Up/Down to adjust vertical viewing angle.
- Look Left/Right to pan horizontally.
- Enjoy an immersive experience with navigating the 360-degree videos in the playlist.
Contributions are welcome! Here's how you can contribute:
- Fork the repository.
- Create a new branch (
git checkout -b feature/improvement
). - Make your changes.
- Commit your changes (
git commit -am 'Add feature/improvement'
). - Push to the branch (
git push origin feature/improvement
). - Create a new Pull Request.
This project is licensed under the MIT License - see the LICENSE file for details.