hello guys i did an ar web app by learning hiukim yuen course and i made both a 3d and video projects with image tracking using mind ar now i have a problem i want that when i load my 3d model i want to have touch event handling something like model viewer but like I want the image tracking like the target be in camera then the 3d model loads and then the user can rotate and do other stuff with it
I was wondering if there are any courses, odcs, or github repositories that could assist me with doing this with my 3D models.
When I was working on a similar AR project, I faced the challenge of integrating touch event handling with 3D models. I found that using a combination of resources helped a lot. I started with tutorials on AR frameworks like AR.js or Vuforia for image tracking, then moved on to more advanced topics with Three.js for 3D model interactions. I also explored GitHub repositories focused on AR and 3D model handling—many have examples of touch and interaction events. For detailed guidance, the Mozilla Developer Network (MDN) and A-Frame documentation provided valuable insights. It was a mix of learning from courses, experimenting with code, and consulting open-source projects that really helped me get the touch interactions working smoothly.
It sounds like you’re looking for guidance on integrating touch event handling with your 3D model in an AR web app. Courses or GitHub repositories focusing on AR development with libraries like Three.js or A-Frame might help. Check platforms like Udemy or GitHub for relevant resources and tutorials.