Visual Pick and Place

From Computer Laboratory Group Design Projects
Revision as of 16:48, 13 November 2018 by afb21 (talk | contribs)
Jump to navigationJump to search

Client: Theo Markettos, Computer Lab <atm26@cam.ac.uk>

We have a LitePlacer robot that can automatically assemble circuit boards (https://youtu.be/t__ybwOufyg), with a machine vision system based on OpenPnP. Unlike many computer vision problems, the cameras could in principle be moved around, getting better field of view or details of specific regions. Our robot currently uses the vision capabilities for simple tasks such as recognising board features and component tapes to align components it is placing. We would like to extend the capabilities of the system to automate the setup of the machine, for instance printing part numbers from CAD data on sticky labels that the vision system reads; configuring placement of component tapes; recognising parts from their shape or markings to rotate the tape; detecting parts and boards from text printed on them; and guiding the user through setup with augmented image views.