Santa Clara, CA – [May 17, 2019]
Magik Eye today announced it will exhibit at the 2019 Embedded Vision Summit, the premier event for innovators who want to bring visual intelligence to products. The Summit takes place May 20-23, 2019, in Santa Clara, CA. The annual event brings together a global audience of companies developing leading-edge, vision-enabled products, including embedded systems, cloud solutions, and mobile applications. Magik Eye will showcase its family of 3D sensing solutions in Booth #619.
Santa Clara, CA – [May 19, 2019]
Magik Eye today announced it will reveal Invertible Light™ at the 2019 Embedded Vision Summit, the premier event for innovators who want to bring visual intelligence to products. Invertible Light™ is a new method for depth sensing that enables the smallest, fastest and most power-efficient 3D sensing method. “While current methods like Structured Light, Time of Flight and Stereo Scoping Imaging methods are the primary methods today, Invertible Light™ aims to transform 3D sensing in the coming age of robotics and machine vision for the masses,” said Takeo Miyazawa, Founder & CEO of Magik Eye who will be on stage delivering a presentation to reveal Invertible Light at the Summit and to the World.
The Summit takes place May 20-23, 2019, in Santa Clara, CA. The annual event brings together a global audience of companies developing leading-edge, vision-enabled products, including embedded systems, cloud solutions, and mobile applications. Invertible Light™ will be presented by Takeo Miyazawa, Founder & CEO of Magik Eye on May 22nd as part of the Emerging Technologies Track at the Summit.
The ability to see clearly and understand the surrounding environment is critical for robots and many machine vision applications. Current 3D sensors often have trouble with things like reflective surfaces, un-textured white walls, bright sunlight, dark interior spaces, and translucent objects such as windows. MagikEye was designed to see all of that, and more.
Our family of 3D sensing solutions support a wide range of applications, from smartphones, robotics and surveillance applications. Our solutions offer the smallest, fastest and most power-efficient 3D sensing. MagikEye’s technology also includes greater field of view (FOV) flexibility than all other methods. This flexibility includes hemispherical fish eye depth sensing and 90-degree FOV that is capable of speeds up 300+ frames per second (FPS).
Optics & Math Breakthrough enables Smallest,
Fastest & Most Power-Efficient 3D sensing.
Magik Eye’s technology is based on Invertible Light™ which is our new theory for depth sensing. Invertible Light™ is a breakthrough of optics and mathematics that generates a 3D point cloud, using only a regular dot projector and an image sensor. By projecting a regular dot pattern on an object, the Invertible Light™ method, composed of a small set of data and a direct algorithm, computes 3D depth data in an ultra-high speed manner. The result is the smallest, fastest & most power-efficient 3D sensing.
Breakthrough of Optics and Mathematics
Invertible Light™ is in stark contrast to current 3D sensing methods. For example, Structured Light requires the projection of a specific or random pattern to measure distortions. The result is significant power, multiple components and complexity of production. In contrast to the Time of Flight (ToF) method that is good for longer distances, Invertible Light™ has greater visibility in shorter distances with similar advantages of size, speed and power-efficiency.
Regular Pattern Benefits
Less laser power consumption
Variable FOV and lens distortion
Efficient calibration process instead of active alignment
Higher S/N to handle more variety of reflective materials
Direct algorithm with minimal computational power
Can use most optically efficient lasers (1/10th of Structured Light)
Implemented with a VCSEL & DOE and clever algorithm
Simplified production process with no active alignment needed
Mobile devices are changing the way we communicate, shop, and more. Cameras are more than just taking pretty pictures. With Magik Eye, they can recognize individual faces and differentiate them
Businesses and consumers are increasingly depending on robots to perform tasks to support human capability. Magik Eye ensures these powerful robots see the world crystal clear.
The world of mixed reality offers up a new and exciting way to see the world. With Magik Eye, the ability to sense in full 360 degrees will enrich the experience and enhance the viewer’s perception.
Our company was established in 2015 to bring about a new vision.
Utilizing techniques from brain science, we are enabling new types of products that see and understand the world around them to take better action.
See it all. Know it all.
Founder & CEO
Scientific entrepreneur & computer vision expert with a rich history of venture startups worldwide.
Founder, Cognitens, acquired by Hexagon
B.A. Psychology, Sophia University, Tokyo
Researcher in the field of artificial intelligence with an expertise in computer vision, and robotics camera calibration.
Software Engineering, Sun Microsystems
Researcher, Czech Technical University
Ph.D., Computer Vision, Czech Technical University
M.S. Computer Science, Charles University, Prague
Venture capitalist, entrepreneur, and management consultant with a global portfolio
Managing Director, Pound Capital
Vice President of Marketing, Kizna
Management Consultant, PwC
B.A. Political Science, University of Toronto
Head of Japan Operations
High technology entrepreneur and manager of semiconductor ventures.
Founder, CTO of NuCore acquired by Mediatek
Founder, CEO of LivingImage
Led engineering & operations Gentech, Hitachi Medical & Intel
B.S. Computer Science, Keio University
Professor, Department of Brain & Cognitive Sciences, MIT
Director, Center for Brains, Minds, and Machines
Co-Director, Center for Biological and Computational Learning
Director, Mobile Eye and early Advisor to Deep Mind
Professor, Department of Computer Science, Columbia University
Director, Columbia University Machine Learning Laboratory
Director, Machine Learning, Netflix
Assistant Professor, Czech Technical University, Department of Cybernetics
Distinguished Researcher for Center of Machine Perception
We are looking for new colleagues for our Prague or Tokyo office
Initiate development. Be an important part of the design and development for the depth sensing module.
Work in close collaboration with the Prague and Japan teams on the sensor's design and programming.
Bachelor or Master degree in a relevant field (Computer Science, IT Engineering, or other Engineering studies, depending on experience)
Programming experience in C++ and Python
Experience with software project management (cmake, git, ...)
Interest in cameras and depth capturing techniques
Fluent English proficient
Nice to have
Knowledge of computer hardware and electronics design
Experience with Windows, Linux, and network administration
Computer Vision experience
Testing of mechanical devices under various conditions
Designing test cases
Reporting of results and problem solving
Experience with data acquisition (i.e. camera, Kinect, LIDAR)
Basic scripting knowledge (Python or Matlab)
English language proficient
Nice to have
Experience with administration of Windows/Linux/Embedded systems
Experience with industrial robots
Computer Vision experience
1055 Washington Boulevard, Stamford, CT, 06901
CIIRC, Jugoslávských partyzánů 1580/3 160 00 Praha 6, Czech Republic
ARK LAND Yotsuya 5F
15-5 Yotsuya-saneicho, shinjuku-ku, Tokyo, Japan