Object detection technology: The next big thing
Seagate Technology inspects tens of millions of computer storage drive parts daily. The media that goes into them. The heads. The wafers. The completed assemblies and outer shells.
For much of the company's history, this inspection was a time-consuming and costly process. To achieve the quality the global company’s customers expect, hard disk and solid-state drives would be tested once, twice, or even three times before human beings gave them a final blessing in what’s known as sequential inspections, which involves looking intently at parts through electron microscopes.
But with the rise of AI-enabled object detection technology, Seagate is automating inefficiency out of its production processes and bringing in high-speed consistency. To achieve this increased throughput, the company is now installing high-powered edge servers closer to the point of production to run its machine learning models and expand its Fourth Industrial Revolution technology to assembly lines in factories around the world.
AI technologies at the edge
“Our quality systems use a multitude of different types of measurements that are image-based,” says Bruce King, a data science technologist at Seagate. “It’s a rules-based system that is programmed by engineers. The opportunity that we have with artificial intelligence is that we can have the machine learn all the rules so we don’t have to program them. This saves the engineers quite a bit of time and effort. And we’re finding that the machine learning tools are very flexible. They can find all the same things that a rule-based system can. They self-learn, so to speak. And their accuracy is as good or better in all cases that we’ve analyzed so far.”
Seagate is one of hundreds—maybe thousands—of companies that are either planning, piloting, or deploying machine learning with deep neural networks, algorithms patterned after the human brain, to solve real-world business challenges. The good news is that technology techniques like deep learning, and a treasure trove of publicly available frameworks and libraries, have made advanced analytics commercially viable and able to scale. The technologies may vary, and opinions may differ, but like any emerging technologies and standards, arriving at a clear consensus regarding platforms and approaches will continue to evolve. In the meantime, companies like Seagate are getting an early-mover advantage over their competition.
Some of the common debates are around what to sensor, which high-resolution camera to use, which aggregation and data transport software to use, and how to manage and secure the devices at the edge. Organizations must also consider what edge platform is best suited for a majority of their use cases, to minimize complexity and sprawl, as well as the need to hire, versus contract, super-smart data scientists to make object detection work, as demand for this scarce talent pool grows.
Whatever the case, there is clear momentum behind object detection technology. Manufacturers, local governments, transportation authorities, marinas, schools, hospitals, and other vertical industries are increasingly embracing projects built on it.
Innovations at the edge are transforming manufacturing, public safety and more. See what it takes to maximize your data to accelerate your digital business from device to data center.
Factories of the future
In manufacturing, for example, Foxconn, a Taiwanese multinational electronics contract manufacturing company, is about to launch an object detection program to inspect every Hewlett Packard Enterprise server and component it produces for distribution in Europe. Indeed, its Kutná Hora plant in the Czech Republic is designed to inspect multiple products coming down the line and can also examine any variety of configurations rolling along, regardless of their order. The system will examine the objects, compare them with images in the company’s database, and with the help of video analytics and AI, determine whether anything looks amiss. In short, the system enables Foxconn to identify any potential failures in real time, which is critical given the company’s efforts to ship orders to customers like HPE as rapidly as possible, while ensuring consistent high quality across the products produced.
“The system compares what’s been assembled to what it believes should be assembled,” says John Gallagher, operations manager at Foxconn CZ. “It will look to see if everything is where it should be. Are the processors in the right spots? Are the cables in the right sockets? Are the drives correct? And it will go through the whole configuration that we ask it to identify. The benefit of all of this to us is that we try to ship 50 percent of the orders we receive within 24 hours. This [object detection technology] allows us to be more predictable for customers. It allows us to have shorter manufacturing times for them as well. And it helps remove failures from the process.”
Gallagher says the company has high hopes for the pilot project, which kicked off in late March. If all goes well, Foxconn could expand the technology to other continents.
Object detection technology is also showing great promise in the world of transportation, most notably with how people park their cars and moor their boats.
If you drive to work, you may dread the usual slog of trying to find parking on a city street or in a garage. In the near future, this might not be so difficult, thanks to smart parking solutions.
Instead of driving around for 10, 15, or even 30 minutes trying to find a parking space, you whip out a mobile app that locates a parking spot for you. On the back end, some combination of IoT parking meters, sensors, video cameras, radar, analytics software, DNN programs, edge servers, and mobile payment systems might come together to find that lone spot at the corner of 5th and Market—and even reserve it for you. In fact, you might eventually be able to sign up for an AI-driven service that knows your typical driving habits—when you go to work, when you stop off at the neighborhood bar, and when you go home—and prearranges everything before you even think to do it yourself.
Sure, that’s all pretty futuristic. But the technology is evolving so quickly that it’s not beyond the realm of possibility. And we are already seeing startups and local governments as well as transportation authorities working together to try and make something happen before too long.
In fact, one research firm says the smart parking technologies market is expected to eclipse $5.51 billion by 2024, with a compound annual growth rate of 19 percent.
“The provision and management of parking within cities is critical for the city’s ability to function and thrive,” notes Mark Zannoni, IDC research director for smart cities and transportation. "Parking impacts the urban economy and quality of life, as it provides access to businesses and events, affects traffic, and provides sizable revenues to local governments. Accordingly, more and more cities worldwide are looking at smart parking as a service offering while they transform themselves into smart cities."
Los Angeles could be considered one such smart city. Since 2012, its LA Express Park program has sought to make more parking available in a 4.5-square-mile downtown area by giving commuters several ways to find parking spots. This has always been a huge issue for the City of Angels, where much of downtown peak traffic congestion is thought to be caused by drivers hunting for parking.
Through LA Express and work being done with a San Francisco Bay Area startup called Streetline, the city has slowly expanded the use of smart parking meters, parking space vehicle sensors, real-time parking guidance systems, and an integrated parking management system to other districts. Streetline has partnered with other California cities, including San Carlos and San Mateo, to save people nearly 713,000 hours of driving, 3 million miles, and 178,00 gallons of gas, according to its website.
“We are still in the position that hard, real data is required to power real-time availability of parking information,” says Mark Noworolski, CTO at Streetline. “Connected cameras with an object detection back end are a natural fit for that, as they are less invasive than on-street sensors and can be deployed in a wide variety of environments.”
The same technology that may save you time on the road could also help you accurately anchor your boat. At the recent Dusseldorf Boat Show in Germany, FLIR Systems, a Wilsonville, Oregon-based imaging and surveillance systems company, announced what may be the marine industry’s first intelligent object recognition and motion sensing “assisted docking solution.” Called Raymarine DocSense, the company’s vision camera technology and video analytics integrate intelligence gathered from surrounding imagery with the vessel’s propulsion and steering system to assist boat owners in tight-quarter docking maneuvering.
One of the more profound and meaningful applications of object detection technology will be how it helps protect children.
Since the Columbine school massacre in 1999, more than 187,000 students have been exposed to gun violence, according to a Washington Post study. It’s a tragedy that continues to grow and intensify with few good solutions. For example, some suggest deploying armed guards while others advocate hoisting security gates on campuses. Neither approach has much support.
But what if you could almost invisibly identify whether guns, knives, or other weapons—concealed or visible—are coming into schools? If you could do so without violating privacy rights, wouldn’t it be interesting? And wouldn’t it be worth discussing possibly expanding such technology to other threatened areas, such as airports, bus stations, hotels, casinos, movie theaters and shopping malls?
Numerous startups not only believe it’s possible but that they have a calling to do so.
“This is a relatively new field. But given the urgency of the [mass shooting] problem, there are a number of players looking at how to address the challenge,” says Martin Cronin, CEO of Canada’s Patriot One Technologies, which offers a concealed weapons detection system built around a cognitive microwave radar device and custom AI software. “There is no question that embedded, discreet sensors, powered by AI, will become more ubiquitous for public safety.”
Cronin acknowledges that any discussion about monitoring people to see what they’re carrying—even if it’s guns—is always controversial. For that reason, he says it is imperative to involve the public in all decisions around security measures.
“There are understandably public fears about slipping quietly into a mass surveillance society,” Cronin says. “Our technologies focus on the threat objects (guns, knives, bombs, etc.). The person is of no interest to the system unless they are carrying a threat. That is important to maintaining public confidence and for assuring that there is no discrimination involved. We also generate no body image in the process of scanning for weapons with our radar-based systems, which is important for privacy. This is an approach which we feel can maintain the important public consensus. People want to feel safe, but there is a limit to what they are prepared to give up in their civil liberties to enjoy that safety.”
Object recognition technology is in its early stages. While it has begun to transform the manufacturing floor because the system works with known objects, the long-term vision for the technology is to infer things about objects and learn from those experiences. AI and machine learning techniques using multilayered applications of deep neural networks will help it to eventually be able to pick out almost any object in a crowd. But it’s not quite there yet. A team of cognitive psychologists from the University of California, Los Angeles, challenged some of the best deep learning networks to look at color images of animals and objects, with the images altered. For example, the surface of a golf ball was displayed on a teapot; zebra stripes were placed on a camel; and the pattern of a blue and red argyle sock was shown on an elephant. The learning networks ranked their top choices and chose the correct item as the first choice for only five of 40 objects.
“Deep learning systems process information very differently from humans,” notes Philip Kellman, a UCLA distinguished professor of psychology and senior author of the study. “The biggest misunderstanding is that AI devices are 'seeing' the way we are. That's not a comment about having consciousness. When a deep network is shown a picture and says ‘horse,’ it does not have a description of what the shape of the object is [and] does not perceive that it is detachable from the surfaces behind it. Conversely, if you view a certain glass ornament on my desk, you will immediately see that it is a goose, because you see its shape. Deep learning systems aren't seeing that shape at all. And shape is not the only abstract property that's being missed—but it is among the most crucial. In human vision and learning, the fascinating fact is that you don't need to exhaustively show all possible views and versions of the objects that you want the network to learn because when we acquire abstract descriptions, like shape, we have a more powerful way of generalizing.”
Why does that matter?
“We want these systems to work reliably,” says Hongjing Lu, a UCLA professor of psychology and the author of another study. “If a visual system used for self-driving cars failed to recognize a stop sign because of some graffiti on it, that would create a serious safety issue. Cybersecurity is also a big concern if AI systems can be easily fooled by fake images and videos.”
There are, however, some signs that the technology will become more human-like over time.
Vwani Roychowdhury, a UCLA professor of electrical and computer engineering, likens the current deep learning systems to an efficient memorization engine.
“The premise,” Roychowdhury says, “is that if you train these deep networks with large enough datasets, then the surprise factor can be mitigated: Whenever it is shown a new image of a car or a human, it will be very similar in shape, texture, and other visual attributes to something that it has already seen before during the training or the memorization phase. But in reality, for most situations and objects, such exhaustive training is just not feasible; there are just too many combinations to memorize.”
Motivated by the challenge, UCLA and Stanford University engineers working with Roychowdhury demonstrated a computer system that can discover and identify the real-world objects it “sees” based on the same method of visual learning humans use. The approach was to help the systems “learn” like humans by immersing them in an Internet replica of environments humans experience every day. Like humans, their system automatically builds shape- and parts-based models for individual objects from the wealth of images and videos that depicts them from many perspectives and in different settings. Thus, it circumvents the need for exhaustive memorization and recognizes objects in images it has never seen before—also like humans.
Object detection technology is in the process of proving itself, and much like the original Industrial Revolution, it has the potential, at the very least, to free people from tedious jobs that will be done more efficiently and effectively by machines. It will also open up new avenues of research and operations that will reap additional benefits.
While there are still challenges in applying object detection to more nuanced tasks, with its continued evolution, along with the devices and techniques that make it possible, it could soon become the next big thing.
Object detection: Lessons for leaders
- Object detection in manufacturing builds on existing automation technologies.
- The technology can streamline common, everyday activities such as parking flow and traffic management.
- Reliability is the key to successful deployments.
This article/content was written by the individual writer identified and does not necessarily reflect the view of Hewlett Packard Enterprise Company.