Hacker News new | past | comments | ask | show | jobs | submit login

I've played around with both three.js and babylon.js, in attempts to make a CAD system. Where I get in over my head is moving objects relative to each other. For example, I need cube A's top right front corner to be 10mm left of the midpoint of cube B's rear face, left edge. Fusion360 can select any of these edges, vertices, etc, and gives you the controls and necessary visual feedback, with distance, angles and alignments.

Trying to implement even the simplest version of this has been a project killer.




I don't understand why this cannot be done , One just had to calculate the vector and displace cube by said amount . I think I'm missing something ?


I'm not talking about the user writing code to move an object. I'm talking about the user clicking features on 3D objects with their mouse (like a specific vertex), having those features become highlighted, drawing the visual cues that let the user know that they are in "Move" mode, showing distances and angles to give the user the necessary feedback, clipping to features like midpoints, etc.

Really, the piece that drove me to eventually give up was getting a vertex or edge to be selected, rather than the whole object. Its not difficult to calculate for simple shapes, but what is really needed is a general solution, that doesn't depend on shapes being regular polygons.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: