Sign up to save tools and stay up to date with the latest in AI
bg
bg
1

GitHub - collidingScopes/3d-model-playground: Control 3D models using hand gestures and voice commands in real-time. Threejs / mediapipe computer vision

Jun 03, 2025 - github.com
The "3D Model Playground" is an interactive web application that allows users to control 3D models using hand gestures and voice commands in real-time. Built with technologies like Three.js, MediaPipe, Web Speech API, and Rosebud AI, the app enables users to interact with 3D models by saying commands such as "drag," "rotate," "scale," or "animate," and using finger pinches for control. Users can also import new 3D models in GLTF format by dragging and dropping them onto the page. The application requires a modern web browser with WebGL support and access to a camera and microphone.

For development, users can clone the repository from GitHub and serve it locally using Python's HTTP server. The project is open-source under the MIT License and credits various technologies and resources like Three.js, MediaPipe, and Quaternius 3D models. The creator, Alan, shares related projects and contact information, encouraging donations through "Buy Me A Coffee" to support his work in open-source software development.

Key takeaways:

  • 3D Model Playground is an interactive web app that allows users to control 3D models using hand gestures and voice commands in real-time.
  • The app is built using technologies like Three.js, MediaPipe, Web Speech API, and JavaScript for real-time interaction.
  • Users can import new 3D models in GLTF format by dragging and dropping them onto the page.
  • The project is open source and available under the MIT License, with setup instructions provided for development.
View Full Article

Comments (0)

Be the first to comment!