Robohub.org
 

Vision sensor capable of detecting moving spots 0.05mm in size across from distance of 2m

by
26 November 2012



share this:
12-0216-n

Technos has introduced the Super5000K 7K Model neuro-visual sensor. This is the world’s highest precision visual inspection system, and can detect spots 0.05mm in size at a distance of 2m, with a 1.4m field of vision. This is 1,000 times the resolving power of a conventional 4,000 pixel line sensor CCD camera, and 4,000 times that of a full HD camera.

“For industrial applications, this product has recently been used in many automotive-related situations, but it is also used in the steel, electronics, and flat-panel display industries. Typical manufacturing processes end with a visual inspection; we can fully automate that. Our sensor has high speed and high precision. It operates on the same principles as the human eye, but with 100 times the precision. It can perform automatic inspections with 1,000 times the precision of conventional CCDs. In principle, it emulates the movement of cells in the human eye; your eye vibrates up and down 80 times per second, and we emulate that vibration using electronic circuits.”

Conventional CCD-based equipment has had difficulty detecting color variations. By applying the principles of human vision, Technos has achieved a sensor with 100 times the precision of human eyes, making it possible to detect color variation.

“Color variation is a problem in a variety of settings; this sensor is used in industrial applications, but is also used in maintenance applications such as inspecting highways or oil storage tanks. This technology for rapidly picking up small details will be even more widely used in the future, and we plan to develop those applications.”

Pricing for the minimum configuration starts at an equivalent of $240,000, and runs up to between $470,000 and $730,000, depending on the particular specifications. Technos estimates that about 240 companies listed on the first section of the Tokyo Stock Exchange will need equipment with this level of precision. The company aims to sell about 10 units per year to the automobile, steel, semiconductor, and liquid crystal industries. Technos has been awarded patents in 14 countries around the world. With inquiries coming from foreign countries, it is looking to expand overseas as well.



tags: ,


DigInfo TV is a Tokyo-based online video news platform dedicated to producing original coverage of cutting edge technology, research and products from Japan.
DigInfo TV is a Tokyo-based online video news platform dedicated to producing original coverage of cutting edge technology, research and products from Japan.





Related posts :



Interview with Dautzenberg Roman: #IROS2023 Best Paper Award on Mobile Manipulation sponsored by OMRON Sinic X Corp.

The award-winning author describe their work on an aerial robot which can exert large forces onto walls.
19 November 2023, by

Robot Talk Episode 62 – Jorvon Moss

In the latest episode of the Robot Talk podcast, Claire chatted to Jorvon (Odd-Jayy) Moss from Digikey about making robots at home, and robot design and aesthetics.
17 November 2023, by

California is the robotics capital of the world

In California, robotics technology is a small fish in a much bigger technology pond, and that tends to conceal how important Californian companies are to the robotics revolution.
12 November 2023, by

Robot Talk Episode 61 – Masoumeh Mansouri

In the latest episode of the Robot Talk podcast, Claire chatted to Masoumeh (Iran) Mansouri from the University of Birmingham about culturally sensitive robots and planning in complex environments.
10 November 2023, by

The 5 levels of Sustainable Robotics

Robots can solve the UN SDGs and not just via the application area.
08 November 2023, by

Using language to give robots a better grasp of an open-ended world

By blending 2D images with foundation models to build 3D feature fields, a new MIT method helps robots understand and manipulate nearby objects with open-ended language prompts.
06 November 2023, by





©2021 - ROBOTS Association


 












©2021 - ROBOTS Association