It was a return to the source for RoboCup 2017, which took place last week in Nagoya Japan, 20 years after its launch in the same city.
Bigger than ever, the competition brought together roboticists from around the world. Originally focussed on robot football matches, RoboCup has expanded to include leagues for rescue robots, industrial robots, and robots in the home. Kids are also part of the fun, competing in their own matches and creative shows. You can watch video introductions of all of the leagues here, or watch a quick summary below.
And here’s a scroll through 5 hours of football glory from this year’s competition.
The data captured by today’s digital cameras is often treated as the raw material of a final image. Before uploading pictures to social networking sites, even casual cellphone photographers might spend a minute or two balancing color and tuning contrast, with one of the many popular image-processing programs now available.
This week at Siggraph, the premier digital graphics conference, researchers from MIT’s Computer Science and Artificial Intelligence Laboratory and Google are presenting a new system that can automatically retouch images in the style of a professional photographer. It’s so energy-efficient, however, that it can run on a cellphone, and it’s so fast that it can display retouched images in real-time, so that the photographer can see the final version of the image while still framing the shot.
The same system can also speed up existing image-processing algorithms. In tests involving a new Google algorithm for producing high-dynamic-range images, which capture subtleties of color lost in standard digital images, the new system produced results that were visually indistinguishable from those of the algorithm in about one-tenth the time — again, fast enough for real-time display.
The system is a machine-learning system, meaning that it learns to perform tasks by analyzing training data; in this case, for each new task it learned, it was trained on thousands of pairs of images, raw and retouched.
The work builds on an earlier project from the MIT researchers, in which a cellphone would send a low-resolution version of an image to a web server. The server would send back a “transform recipe” that could be used to retouch the high-resolution version of the image on the phone, reducing bandwidth consumption.
“Google heard about the work I’d done on the transform recipe,” says Michaël Gharbi, an MIT graduate student in electrical engineering and computer science and first author on both papers. “They themselves did a follow-up on that, so we met and merged the two approaches. The idea was to do everything we were doing before but, instead of having to process everything on the cloud, to learn it. And the first goal of learning it was to speed it up.”
Short cuts
In the new work, the bulk of the image processing is performed on a low-resolution image, which drastically reduces time and energy consumption. But this introduces a new difficulty, because the color values of the individual pixels in the high-res image have to be inferred from the much coarser output of the machine-learning system.
In the past, researchers have attempted to use machine learning to learn how to “upsample” a low-res image, or increase its resolution by guessing the values of the omitted pixels. During training, the input to the system is a low-res image, and the output is a high-res image. But this doesn’t work well in practice; the low-res image just leaves out too much data.
Gharbi and his colleagues — MIT professor of electrical engineering and computer science Frédo Durand and Jiawen Chen, Jon Barron, and Sam Hasinoff of Google — address this problem with two clever tricks. The first is that the output of their machine-learning system is not an image; rather, it’s a set of simple formulae for modifying the colors of image pixels. During training, the performance of the system is judged according to how well the output formulae, when applied to the original image, approximate the retouched version.
Taking bearings
The second trick is a technique for determining how to apply those formulae to individual pixels in the high-res image. The output of the researchers’ system is a three-dimensional grid, 16 by 16 by 8. The 16-by-16 faces of the grid correspond to pixel locations in the source image; the eight layers stacked on top of them correspond to different pixel intensities. Each cell of the grid contains formulae that determine modifications of the color values of the source images.
That means that each cell of one of the grid’s 16-by-16 faces has to stand in for thousands of pixels in the high-res image. But suppose that each set of formulae corresponds to a single location at the center of its cell. Then any given high-res pixel falls within a square defined by four sets of formulae.
Roughly speaking, the modification of that pixel’s color value is a combination of the formulae at the square’s corners, weighted according to distance. A similar weighting occurs in the third dimension of the grid, the one corresponding to pixel intensity.
The researchers trained their system on a data set created by Durand’s group and Adobe Systems, the creators of Photoshop. The data set includes 5,000 images, each retouched by five different photographers. They also trained their system on thousands of pairs of images produced by the application of particular image-processing algorithms, such as the one for creating high-dynamic-range (HDR) images. The software for performing each modification takes up about as much space in memory as a single digital photo, so in principle, a cellphone could be equipped to process images in a range of styles.
Finally, the researchers compared their system’s performance to that of a machine-learning system that processed images at full resolution rather than low resolution. During processing, the full-res version needed about 12 gigabytes of memory to execute its operations; the researchers’ version needed about 100 megabytes, or one-hundredth as much. The full-resolution version of the HDR system took about 10 times as long to produce an image as the original algorithm, or 100 times as long as the researchers’ system.
“This technology has the potential to be very useful for real-time image enhancement on mobile platforms,” says Barron. “Using machine learning for computational photography is an exciting prospect but is limited by the severe computational and power constraints of mobile phones. This paper may provide us with a way to sidestep these issues and produce new, compelling, real-time photographic experiences without draining your battery or giving you a laggy viewfinder experience.”
Artificial skin with post-human sensing capabilities, and a better understanding of skin tissue, could pave the way for robots that can feel, smart-transplants and even cyborgs.
Few people would immediately recognise the skin as our bodies’ largest organ, but the adult human has on average two square metres of it. It’s also one of the most important organs and is full of nerve endings that provide us with instant reports of temperature, pressure and pain.
So far the best attempts to copy this remarkable organ have resulted in experimental skin with sensor arrays that, at best, can only measure one particular stimulus.
But the SmartCore project, funded by the EU’s European Research Council and at the Graz University of Technology (TU Graz) in Austria, hopes to create a material that responds to multiple stimuli. To do so requires working at a nanoscale — where one nanometre represents a billionth of a metre — creating embedded arrays of minuscule sensors that could be 2 000 times more sensitive than human skin.
Principal investigator Dr Anna Maria Coclite, an assistant professor at TU Graz’s Institute for Solid State Physics, says the project aims to create a nanoscale sensor which can pick up temperature, humidity and pressure — not separately, but as an all-in-one package.
‘They will be made of a smart polymer core which expands depending on the humidity and temperature, and a piezoelectric shell, which produces an electric current when pressure is applied,’ she said.
These smart cores would be sandwiched between two similarly tiny nanoscale grids of electrodes which sense the electrical charges given off when the sensors ‘feel’ and then transmit this data.
If the team can surmount the primary challenge of distinguishing between the different senses, the first prototype should be ready in 2019, opening the door for a range of test uses.
Robots
Dr Coclite says the first applications of a successful prototype would be in robotics since the artificial skin they’re developing has little in common with our fleshy exterior apart from its ability to sense.
‘The idea is that it could be used in ways, like robotic hands, that are able to sense temperatures,’ said Dr Coclite. ‘Or even things that can be sensed on even a much smaller scale than humans can feel, i.e, robotic hands covered in such an artificial skin material that is able to sense bacteria.’
Moreover, she says the polymers used to create smart cores are so flexible that a successful sensor could potentially be modified in the future to sense other things like the acidity of sweat, which could be integrated into smart clothes that monitor your health while you’re working out.
And perhaps, one day, those who have lost a limb or suffered burns could also benefit from such multi-stimuli sensing capabilities in the form of a convincingly human artificial skin.
‘It would be fantastic if we could apply it to humans, but there’s still lots of work that needs to be done by scientists in turning electronic pulses into signals that could be sent to the brain and recognised,’ said Dr Coclite.
She also says that even once a successful prototype is developed, possible cyborg use in humans would be at least a decade away — especially taking into account the need to test for things like toxicity and how human bodies might accept or reject such materials.
Getting a grip
But before any such solutions are possible, we must learn more about biological tissue mechanics, says Professor Michel Destrade, host scientist of the EU-backed SOFT-TISSUES project, funded by the EU’s Marie Skłodowska-Curie actions.
Prof. Destrade, an applied mathematician at the National University of Ireland Galway, is supporting Marie Skłodowska-Curie fellow Dr Valentina Balbi in developing mathematical models that explain how soft tissue like eyes, brains and skin behave.
‘For example, skin has some very marked mechanical properties,’ said Prof. Destrade. ‘In particular its stretch in the body — sometimes you get a very small cut and it opens up like a ripe fruit.’
This is something he has previously researched with acoustic testing, which uses non-destructive sound waves to investigate tissue structure, instead of chopping up organs for experimentation.
And in SOFT-TISSUES’ skin research, the team hopes to use sound waves and modelling as a cheap and immediate means of finding the tension of skin at any given part of the body for any given person.
‘This is really important to surgeons, who need to know in which direction they should cut skin to avoid extensive scarring,’ explained Prof. Destrade. ‘But also for the people creating artificial skin to know how to deal with mismatches in tension when they connect it to real skin.
‘If you are someone looking to create artificial skin and stretch it onto the body, then you need to know which is the best way to cut and stretch it, the direction of the fibres needed to support it and so on.‘
Dr Balbi reports that the biomedical industry has a real hunger for knowledge provided by mathematical modelling of soft tissues — and especially for use in bioengineering.
She says such knowledge could be useful in areas like cancer research into brain tumour growth and could even help improve the structure of lab-grown human skin as an alternative to donor grafts.
Sixteen teams from across the globe came to Nagoya, Japan to participate in the third annual Amazon Robotics Challenge. Amazon sponsors the event to strengthen ties between the industrial and academic robotics communities and to promote shared and open solutions to some of the big puzzles in the field. The teams took home $270,000 in prizes.
Watch the video below to see these inventive teams – and their robots – in action.
Congratulations to this year’s winners from the Australian Centre for Robotic Vision. See the full results here.
July 2017 was a big month for robotics-related company funding. Four raised $588 million and 19 others raised $370.6 million for a monthly total of $958.6 million. Acquisitions also continued to be significant with ST Engineering acquiring Aethon for $36 million, iRobot buying its European distributor for $141 million, and SoftBank purchasing 5% of iRobot shares for around $120 million.
Fundings
Plenty, a San Francisco vertical farm startup, raised $200 million in a Series B round led by SoftBank and included Bezos Expeditions, Data Collective, DCM Ventures, Finistere Ventures, Innovation Endeavors and Louis Bacon. Plenty plans to use the funds to expand to Japan and add strawberries and cucumbers to the leafy greens they already produce. Plenty makes an internet-connected system which delivers specific types of light, air composition, humidity and nutrition, depending on which crop is being grown, and is designing and adding robotics and automation as it can, particularly with their recent acquisition of Bright Agrotech (see below). Plenty says it can yield up to 350 times more produce in a given area than conventional farms — with 1 percent of the water.
Sanjeev Krishnan of S2G Ventures said: “This investment shows the potential of the sector. Indoor agriculture is a real toolkit for the produce industry. There is no winner takes all potential here. I could even see some traditional, outdoor growers do indoor ag as a way to manage some of the fundamental issues of the produce industry: agronomy, logistics costs, shrinkage, freshness, seasonality and manage inventory cycles better. There are many different models that could work and we are excited about the platforms being built in the market.”
Nauto, a Silicon Valley self-driving device and AI startup, raised $159 million in a Series B funding round led by SoftBank and Greylock Partners and also included previous investors BMW iVentures, General Motors Ventures, Toyota AI Ventures, Allianz Group, Playground Global and Draper Nexus.
SoftBank Group Corp. Chairman and CEO Masayoshi Son said, “While building an increasingly intelligent telematics business, Nauto is also generating a highly valuable dataset for autonomous driving, at massive scale. This data will help accelerate the development and adoption of safe, effective self-driving technology.”
Desktop Metal, the MIT spin-off and Massachusetts-based 3D metal printing technology startup, raised another $115 million in a Series D round which included New Enterprise Associates, GV (Google Ventures), GE Ventures, Future Fund and Techtronic Industries which owns Hoover U.S. and Dirt Devil.
According to CEO Ric Fulop, “You don’t need tooling. You can make short runs of production with basically no tooling costs. You can change your design and iterate very fast. You can make shapes you couldn’t make any other way, so now you can lightweight a part and work with alloys that are very, very hard, with very extreme properties. One of the benefits for this technology for robotics is that you’re able to do lots of turns. Unless you’re iRobot with the Roomba, you’re making a lot of one-off changes to your product.”
Brain Corp, a San Diego AI company developing self-driving technology, got $114 million in a Series C funding round led by the SoftBank Vision Fund. Qualcomm Ventures was the only other investor. The funds will be used to develop technology that enables robots to navigate in complex physical spaces. Last October, Brain Corp. rolled out its first commercial product —a self-driving commercial floor scrubber for use in grocery stores and big box retailers.
Beijing Geekplus Technology (Geek+), a Chinese startup developing a goods-to-man warehousing system of robots and software very similar to Kiva System’s products, raised $60 million in a B round led by Warburg Pincus and joined by existing shareholders and Volcanics Venture. The company claims to have delivered the largest numbers of logistics robots among its peers in China, delivering nearly 1,000 units of robots in warehouses for over 20 customers that include Tmall, VIPShop and Suning.
Yong Zheng, Founder and CEO of Geek+, said, “This round of financing will help us upgrade our business in three aspects. Firstly, we will accelerate the upgrading of our logistics robotics products and expand product offerings to cover more applications.” “Secondly, we will accelerate our geographical expansion and industry coverage to provide our one-stop intelligent logistics system and operation solutions to more customers. Thirdly, we will start exploring overseas markets through multiple channels.”
Vicarious, a Union City, California-based artificial intelligence company using computational neuroscience to build better machine learning models that help robots quickly address a wide variety of tasks, raised $50 million funding led by Khosla Ventures.
Momenta.ai, a Beijing autonomous driving startup that is developing digital maps, driving decision solutions and machine vision technology to detect traffic signs, pedestrians and track other cars, raised $46 million in a Series B funding round led by NIO Capital. Sequoia Capital China and Hillhouse Capital along with Daimler AG, Shunwei Capital, Sinovation Ventures and Unity Ventures also participated.
Autotalk, an Israeli chip maker of vehicle to vehicle communications, raised $40 million from Toyota, Sumitomo Mitsui Banking and other investors. The funding will allow Autotalks to prepare and expand its operations for the upcoming start of mass productions as well as continue to develop communication solutions for both connected and autonomous cars.
Flashhold (also named Shanghai Express Warehouse Intelligent Technology and Quicktron) raised $29 million in a Series B round led by Alibaba Group's Cainiao Network and SB China Venture Capital (SBCVC). Flashhold is a Shanghai-based logistic robotics company with robotic products, shelving and software very similar to Amazon's Kiva Systems.
Slamtec, a Chinese company developing a solid state LiDAR laser sensor for robots in auto localization and navigation, raised $22 million from Chinese Academy of Sciences Holdings, ChinaEquity Group Inc. and Shenzhen Guozhong Venture Capital Management Co.
6 River Systems, the Boston, MA startup providing alternative fulfillment solutions for e-commerce distribution centers, raised $15 million in a round led by Norwest Venture Partners with participation from Eclipse Ventures and other existing investors.
Prospera, an Israeli ag startup, raised $15 million in a Series B round for its end-to-end internet of things platform for indoor and outdoor farms. The round was led Qualcomm Ventures and fellow telecom heavyweight Cisco. Propsera uses computer vision, machine learning, and data science to detect and identify diseases, nutrient deficiencies, and other types of crop stress on farms with the hope of improving crop yields and saving farmer costs.
“Receiving funding from these major tech companies is a clear signal that tech industry heavy-hitters understand that agriculture is ripe for digitalization. It means that such companies, which are already involved in digitizing other traditional industries, see a significant opportunity in agtech,” said Prospera CEO Daniel Koppel.
Embark, a Belmont, California-based self-driving trucking startup, raised $15 million in Series A funding led by Data Collective and was joined by YC Continuity, Maven Ventures and SV Angel. Embark has teamed up with Peterbilt and plans to hire for their engineering team and add more trucks to expand their test fleet across the U.S.
Xometry, a Maryland startup with an Uber-like system for parts manufacture, raised $15 million in funding led by BMW Groups’ VC arm and GE.
Intuition Robotics, an Israeli startup developing social companion technologies for seniors, raised $14 million in a Series A round led by Toyota Research Institute plus OurCrowd and iRobot as well as existing seed investors Maniv Mobility, Terra Venture Partners, Bloomberg Beta and private investors.
Dr. Gill Pratt, CEO of Toyota Research Institute said: “We are impressed with Intuition Robotics’ thought leadership of a multi-disciplinary approach towards a compelling product offering for older adults including: Human-Robot-Interaction, cloud robotics, machine learning, and design. Specifically, we believe Intuition Robotics’ technology, in the field of cognitive computing, has strong potential to positively impact the world’s aging population with a proactive, truly autonomous agent that’s deployed in their social robot, ElliQ.”
SkySafe, a San Diego, California-based radio-wave anti-drone device manufacturer, raised $11.5 million in Series A funding, according to TechCrunch. Andreessen Horowitz led the round. SkySafe recently secured DoD contracts to provide counter-drone tech for Navy Seals.
Kuaile Zhihui, a Beijing educational robot startup, has raised around $10 million in a Series A funding round led by Qiming Venture Partners and included GGV Capital and China Capital.
Atlas Dynamics, a Latvian UAS startup, raised $8 million from unnamed institutional and individual investors. Funds will be used to advance the development of its Visual Line of Sight (VLOS) and Beyond Visual Line of Sight (BVLOS) drone-based data solutions, and to build its presence in key markets, including North America.
Reach Robotics, a gaming robots developer, raised $7.5 million in Series A funding led by Korea Investment Partners and IGlobe. Reach has produced and sold an initial run of 500 of its four-legged, crab-like, MekaMon bots. MekaMon fits into an emerging category of smartphone-enabled augmented reality toys like Anki.
UVeye, a New York-based startup that develops automatic vehicle inspection systems, has raised $4.5 million in a seed round led by Ahaka Capital. Israeli angel investors group SeedIL Investment Club also participated. Funds will be used to launch its products and expand to international markets, including China.
Miso Robotics, the Pasadena-based developer of a burger-flipping robot, raised $3.1 million in a funding round led by Acacia Research. Interestingly, Acacia is an agency that licenses patents and also enforces patented technologies.
Metamoto, the Redwood City autonomous driving simulation startup, raised $2 million in seed funding led by Motus Ventures and UL, a strategic investor.
Fastbrick, an Australian brick-laying startup, raised $2 million from Caterpillar with an option to invest a further $8 million subject to shareholder approval. Both companies signed an agreement to collaborate on the development, manufacture, selling and servicing of Fastbrick’s technology mounted on Caterpillar equipment.
Acquisitions
Robopolis SAS, the France-based distributor of iRobot products in Europe, is being acquired by iRobot for $141 million. Last year iRobot, in a similar move to bring their distribution network inhouse, acquired Demand Corp, their distributor for Japan.
Bright Agrotech, a Wyoming provider of vertical farming products, technology and systems, was acquired by Plenty, a vertical farm startup in San Francisco. No financial terms were disclosed. Bright has partnered with small farmers to start and grow indoor farms, providing high-tech growing systems and controls, workflow design, education and software.
Singapore Technologies Engineering Ltd (ST Engineering) has acquired robotics firm Aethon Inc through Vision Technologies Land Systems, Inc. (VTLS), and its wholly-owned subsidiary, VT Robotics, Inc for $36 million. This acquisition will be carried out by way of a merger with VT Robotics, a special purpose vehicle newly incorporated for the proposed transaction. The merger will see Aethon as the surviving entity that will operate as a subsidiary of VTLS, and will be part of the Group’s Land Systems sector.
On the Move Systems, a Canadian penny stock trucking systems provider, is merging with California-based RAD (Robotic Assistance Devices), an integrator of mobile robots for security applications. The merger involves RAD receiving 3.5 million shares of OMVS (around $250k).
IPOs and stock transactions
iRobot, the 27-year-old Massachusetts-based maker of the Roomba, has seen its stock soar from news of a purchase of an undisclosed amount of iRobot stock by SoftBank (or the SoftBank Vision Fund). The purchase is reported to be over $100 million and less than $120 million (5% of the market value).
Almost all robocars use maps to drive. Not the basic maps you find in your phone navigation app, but more detailed maps that help them understand where they are on the road, and where they should go. These maps will include full details of all lane geometries, positions and meaning of all road signs and traffic signals, and also details like the texture of the road or the 3-D shape of objects around it. They may also include potholes, parking spaces and more.
The maps perform two functions. By holding a representation of the road texture or surrounding 3D objects, they let the car figure out exactly where it is on the map without much use of GPS. A car scans the world around it, and looks in the maps to find a location that matches that scan. GPS and other tools help it not have to search the whole world, making this quick and easy.
Google, for example, uses a 2D map of the texture of the road as seen by LIDAR. (The use of LIDAR means the image is the same night and day.) In this map you see the location of things like curbs and lane markers but also all the defects in those lane markers and the road surface itself. Every crack and repair is visible. Just as you, a human being, will know where you are by recognizing things around you, a robocar does the same thing.
Some providers measure things about the 3D world around them. By noting where poles, signs, trees, curbs, buildings and more are, you can also figure out where you are. Road texture is very accurate but fails if the road is covered with fresh snow. (3D objects also change shape in heavy snow.)
Once you find out where you are (the problem called “localization”) you want a map to tell you where the lanes are so you can drive them. That’s a more traditional computer map, though much more detailed than the typical navigation app map.
Some teams hope to get a car to drive without a map. That is possible for simpler tasks like following a road edge or a lane. There you just look for a generic idea of what lane markings or road edges should look like, find them and figure out what the lanes look like and how to stay in the one you want to drive in. This is a way to get a car up and running fast. It is what humans do, most of the time.
Driving without a map means making a map
Most teams try to do more than driving without a map because software good enough to do that is also software good enough to make a map. To drive without a map you must understand the geometry of the road and where you are on it. You must understand even more, like what to do at intersections or off-ramps.
Creating maps is effectively the act of saying, “I will remember what previous cars to drive on this road learned about it, and make use of that the next time a car drives it.”
Put this way it seems crazy not to build and use maps, even with the challenges listed below. Perhaps some day the technology will be so good that it can’t be helped by remembering, but that is not this day.
The big advantages of the map
There are many strong advantages of having the map:
Human beings can review the maps built by software, and correct errors. You don’t need software that understands everything. You can drive a tricky road that software can’t figure out. (You want to keep this to a minimum to control costs and delays, but you don’t want to give it up entirely.)
Even if software does all the map building, you can do it using arbitrary amounts of data and computer power in cloud servers. To drive without a map you can must process the data in real time with low computing resources.
You can take advantage of multiple scans of the road from different lanes and vantage points. You can spot things that moved.
You can make use of data from other sources such as the cities and road authorities themselves.
You can cooperate with other players — even competitors — to make everybody’s understanding of the road better.
One intermediate goal might be to have cars that can drive with only a navigation map, but use more detailed maps in “problem” areas. This is pretty similar, except in database size, with automatic map generation with human input only on the problem areas. If your non-map driving is trustworthy, such that it knows not to try problem areas, you could follow the lower cost approach of “don’t map it until somebody’s car pulled over because it could not handle an area.”
Levels of maps
There are two or three components of the maps people are building, in order to perform the functions above. At the most basic level is something not too far above the navigation maps found in phones. That’s a vector map, except with lane level detail. Such maps know how many lanes there are, and usually what lanes connect to what lanes. For example, they will indicate that to turn right, you can use either of the right two lanes at some intersections.
Usually on top of that will be physical dimensions for these lanes, with their shape and width. The position information may be absolute (ie. GPS coordinates) but in most cases cars are more interested in the position of things relative to one another. It doesn’t matter that you drive exactly the path on the Earth that the lane is in, what matters is that you’re in the right lane relative to the edge of the road. That’s particularly true when you have to deal with re-striping.
Maps will have databases of interesting objects. The most interesting will be traffic signals. It is much easier to decode them if you know exactly where they are in advance. Cars also want to know the geometry of sidewalks and crosswalks to spot where pedestrians will be and what it means if they are there.
Somewhat independent of this are the databases of texture, objects or edges which the car uses to figure out exactly where it is on the map. A car’s main job is “stay in your lane” which means knowing the trajectory of the lane and where you are relative to the lane.
Even those who hope to “drive without a map” still want the basic navigation map, because driving involves not just staying in lanes, but deciding what to do at intersections. You still need to pick a route, just as humans use maps in tools like Waze. The human using Waze still often has the job of figuring out where the lanes are and which one to be in for turns, and how to make the turn, but a map still governs where you will be making turns.
The cost of maps
The main reason people seek to drive without a map is the cost of making maps. Because there is a cost, it means your map only covers the roads you paid to map. If you can only drive at full safety where you have a map, you have limited your driving area. You might say, “Sorry, I can’t go down that road, I don’t have a map.”
This is particularly true if mapping requires human labour. Companies like Google started by sending human driven cars out to drive roads multiple times to gather data for the map. Software builds the first version of the map, and humans review it. This has to be repeated if the roads change.
Maps also are fairly large, so a lot of data must be moved, but storage is cheap, and almost all of it can be moved when cars are parked next to wifi.
To bring down this cost, many companies hope to have ordinary “civilian” drivers go out and gather the sensor data, and to reduce the amount of human labour needed to verify and test the maps.
When the road changes
The second big challenge with maps is the fact that roads get modified. The map no longer matches the road. Fortunately, if the map is detailed enough, that’s quite obvious to the car’s software. The bigger challenge is what to do.
This means that even cars that drive on maps must have some ability to drive when the map is wrong, and even absent. The question is, how much ability?
A surprise change in the road should actually be rare. They happen every day of course as construction crews go out on jobs, but it’s only a surprise to the first car to encounter the change. That very first car will immediately log in the databases that there is a change. If it still drives the road, it will also upload sensor data about the new state of the road. We all see construction zones every day, but how often are the first car even to see that zone?
Most construction zones are scheduled and should not be a surprise even to the first car. Construction crews are far from perfect, so there will still be surprises. In the future, as crews all carry smartphones and have strict instructions to log construction activity with that phone before starting, surprises should become even more rare. In addition, in the interests of safety, the presence of such zones is likely to be shared, even among competitors.
Once a problem zone is spotted, all other cars will know about it. Unmanned cars will probably take a very simple strategy and avoid that section of road if they can, until the map is updated. Why take any risk you don’t need to? Cars with a capable human driver in them may decide they can continue through such zones with the guidance of the passenger. (This does not necessarily mean the passenger taking the controls, but instead just helping the car if it gets confused about things like two sets of lane markings, or unusual cones, or a construction flag crew.)
Nissan has also built a system where the car can ask a remote operations center for such advice, if there is data service at the construction zone. Unmanned cars will probably avoid routes where there could be surprise construction in a place with no data service.
As noted above, several teams are trying to make cars that drive without maps, even in construction zones. Even the cars with maps can still make use of such ability. Even if the car is not quite as safe as it is with a correct map, this will be so rare that the overall safety level can still be acceptable. (Real driving today consists of driving a mix of safer and more dangerous roads after all.) The worst case, which should be very rare, would be a car pulling over, deciding it can’t figure out the road and can’t get help from anybody. A crew in another car would come out to fetch it quickly.
The many players in mapping
This long introduction is there to help understand all the different types of efforts that are going on in mapping and localization. There is lots of variation.
Google/Waymo
The biggest and first player, Google’s car team was founded by people who had worked on Google Streetview. For them the idea of getting cars to scan every road in a region was not daunting — they had done it several times before. Their approach uses detailed texture maps of all roads and surrounding areas. Google is really the world’s #1 map company, so this was a perfect match for them.
Waymo’s maps are not based on Google Maps information, they are much more detailed. Unlike Google Maps, which they give out free to everybody to build on top of, the Waymo maps are proprietary, at least for now.
Navteq/Here
The company with the silly name of “Here” was originally a mapping company named Navteq. It was purchased by Nokia, renamed to “Here” and then sold to a consortium of German automakers. They will thus share their mapping efforts, and also sell the data to other automakers. In addition, the company gets to gather data from a giant fleet of cars from its owners and customers.
Here’s product is called “HD Maps” and it has some similarity to Google’s efforts in scope, but they took a lower cost approach to building them. They build a 3D map of the world using LIDAR. You can find an article about their approach at Here.
TomTom
The Dutch navigation company was already feeling the hurt from the move to phone-based navigation, and with my encouragement, entered the space of self-driving car maps. They have decided to take a “smaller data” approach. Their maps measure the width, not just of the road, but of the space around the road. The width is the distance from what you see looking left and looking right. Sensors will measure the presence of trees, poles, buildings and more to build a profile of the width. That’s enough to figure out where you are on the map, and also where you are on the road.
I’m not sure this is the right approach. It can work, but I don’t think there is a lot of merit in keeping the map small. That’s like betting that bandwidth and storage and computing will be expensive in the future. That’s always been the wrong bet.
MobilEye
MobilEye (now a unit of Intel) has cameras in a lot of cars. They provide the ADAS functions (like adaptive cruise control, and emergency braking) for a lot of OEMs, and they are trying to take a lead in self-driving cars. That’s what pushed their value up to $16B when Intel bought them.
MobilEye wants to leverage all those cars by having them scan the world as they drive, and looks for differences from ME’s compact maps. The maps are very small — just 3D locations of man-made objects around the highway. The location of these objects can be determined by a camera using motion parallax — how things like signs and poles move against the background.
ME believes they can get this data compact enough so that every car with their gear can be uploading updates to maps over the cell network. That way any changes to the road will be reflected in their database quickly, and before they get dramatic, and before a self-driving car gets there.
This is a good plan and the company that does this with the most cars sending data will have an advantage. Like TomTom this makes the bad bet that taking low bandwidth will be an important edge. A more interesting question is how strong the value is in live updates. A fleet that is 10x bigger will discover a change to the road sooner, but is there a big advantage to discovering it in 1 minute vs. 10 minutes?
Tesla
Tesla is one of the few companies hoping to drive without a map, or with a very limited map. A very limited map is more like a phone navigation map — it is not used to localize or plan, but does provide information on specific locations, such as details about off-ramps, or locations of potholes.
Tesla also has an interesting edge because they have many cars out there in production with their autopilot system. That gives them huge volumes of new data every day, though it is the limited data of their cameras. By having customers gather data about the roads, that’s given them a jump up.
Civil Maps
Civil Maps is a VC funded mapping startup. Their plan is to use neural network AI techniques to do a better job at turning image and sensor data into maps. That’s a good, but fairly obvious idea. The real challenge will be just how well they can do it. When it comes to the question of turning the map into a map of lane paths that guide where the vehicle drives, errors can’t be tolerated. If the AI software can’t figure out where the lane is, the software in the car isn’t going to do it either. If successful, the key will be to reduce the amount of human QA done on the maps, not to eliminate it.
Civil Maps publishes technical articles on their web site about their approaches — kudos to them.
DeepMap
DeepMap is another VC funded startup trying to generate a whole map ecosystem. They have not said a lot about their approach, other than they want to use data from production cars rather than having survey fleets do the mapping and re-mapping. That’s hardly a big leap — everybody will use that data if they can get it, and the battle there will partly depend on who has access to the data streams from cars that are out driving with good sensors. We’ll see in the future what other special sauce they want to provide.
Others
Almost every team has some mapping effort. Most teams do not roam a large enough set of roads to have encountered the cost and scaling problems of mapping yet. Only those attempting production cars (like Tesla and Audi) that allow driving without constant supervision had truly needed to deal with a very wide road network. In fact, those planning taxi fleets will not have to cover a wide road network for a number of years.
Most players expect to buy from a provider if they can. While all teams seek competitive edges, this is one sector where the edge is less and the value of cooperation is high. Indeed, the big question in mapping as an industry, is will it become cooperative — as in the case of 3 German automakers co-owning “Here Inc.” or will it become a competitive advantage, with one player making a better product because they have better or cheaper maps?
Infrastructure providers
It seems like a natural for the folks who build and maintain infrastructure to map it. A few things stand in the way of that. Because teams will be trusting the safety of their vehicles to their maps, they need to be very sure about the QA. That means either doing it themselves, or working with a provider whose QA process they can certify.
Working with the thousands of agencies who maintain and build roads is another story. Making all their data consistent and safety critical is a big challenge. Providers will certainly make use of data that infrastructure providers offer, but they will need to do expensive work on it in some cases.
Infrastructure providers can and should work to make sure that “surprises” are very rare. They will never be totally eliminated, but things can be improved. One simple step would be the creation of standardized databases for data on roads and road work. Authorities can pass laws saying that changes to the road can’t be done until they are logged in a smartphone app. This is not a big burden — everybody has smartphones, and those phones know where they are. In fact, smartphones used by contractors can even get smarts to notice that the contractor might be doing work without logging it. Old cheap phones could be stuck in every piece of road maintenance equipment. Those phones would say, “Hmm, I seem to suddenly be parked on a road but there is no construction logged for this area” and alert the workers or a control center.
All new road signs could be also logged by a smartphone app. A law could be made to say, “A road sign is not actually legally in effect until it is logged.” In addition, contractors can face financial penalties for changing roads without logging them. “Fire up the app when you start and end work or you don’t get paid” — that will make it standard pretty quickly.
In episode five of season three we compare and contrast AI and data science, take a listener question about getting started in machine learning, and listen to an interview with Joaquin Quiñonero Candela.
Talking Machines is now working with Midroll to source and organize sponsors for our show. In order find sponsors who are a good fit for us, and of worth to you, we’re surveying our listeners.
If you’d like to help us get a better idea of who makes up the Talking Machines community take the survey at http://podsurvey.com/MACHINES.
If you enjoyed this episode, you may also want to listen to:
Meanwhile, at the Los Angeles Times, Maya Lau writes that a civilian oversight board is pushing Los Angeles Sheriff’s Department to stop flying its drones.
In response to Transport’s report on drone impacts, a coalition of drone manufacturers pressed the government to release the data underpinning its findings. (BBC)
At Shephard News, Richard Thomas looks at how the commercial drone market continues to consolidate.
At an event in Washington, Gen. David Goldfein said that the Air Force needs better artificial intelligence in order to improve intelligence collection. (DefenseTech)
At an Ars Live event, Lisa Ling discussed her role as a drone imagery analyst for the U.S. Air National Guard. (Ars Technica)
Amazon has been granted a patent for a system by which its proposed delivery drones scan a customer’s home upon delivering a product in order to develop product recommendations for future purchases. (CNET)
British firm FlyLogix broke a national record for the longest beyond-line-of-sight drone flight during an 80km operation to inspect structures in the Irish Sea. (The Telegraph)
Rohde & Schwarz, ESG, and Diehl unveiled the Guardion, a counter-drone system. (Jane’s)
Researchers at Moscow Technological Institute have developed a defibrillator drone with a range of up to 50km. (TechCrunch)
The U.S. Army Aviation and Missile Research, Development, and Engineering Center is developing a robotic refueling system for helicopters. (Shephard Media)
India’s Defence Research and Development Organisation has developed an unmanned tank for reconnaissance and mine detection. (Economic Times)
Using hundreds of plastic ducks, researchers at University of Adelaide in Australia have demonstrated that drones are more effective for counting birds than traditional techniques. (New Scientist)
Drones at Work
A team from Queensland University of Technology in Australia is planning to use drones to count koalas as part of a conservation initiative. (Phys.org)
Matagorda County and Wharton County in Texas are acquiring three drones for a range of operations. (The Bay City Tribune)
The Fire Department and Police Department of Orange, Connecticut have acquired a drone for emergency operations. (Milford-Orange Bulletin)
A drone carrying cell phones and other contraband crashed into the yard at the Washington State Prison in Georgia. (Atlanta Journal-Constitution)
North Carolina has adopted a bill that expands drone rules to recreational model aircraft and prohibits drone use near prisons. (Triangle Business Journal)
The U.S. Air Force awarded the University of Arizona a $750,000 grant to build autonomous drones to patrol the U.S. border with Mexico. (Photonics)
The Dallas Safari Club Foundation awarded Delta Waterfowl, a duck hunting organization, a $10,000 grant to use drones to conduct a survey of duck nests. (Grand Forks Herald)
In a statement, Dassault CEO Éric Trappier said that the French-U.K. collaboration on a fighter drone will continue in spite of Brexit and a new Franco-German manned fighter project. (FlightGlobal)
A U.S. military study found that the cost of the Navy’s MQ-4C Triton program has risen by 17 percent. (IHS Jane’s Defense Weekly)
For updates, news, and commentary, follow us on Twitter.
In recent years engineers have been developing new technologies to enable robots and humans to move faster and jump higher. Soft, elastic materials store energy in these devices, which, if released carefully, enable elegant dynamic motions. Robots leap over obstacles and prosthetics empower sprinting. A fundamental challenge remains in developing these technologies. Scientists spend long hours building and testing prototypes that can reliably move in specific ways so that, for example, a robot lands right-side up upon landing a jump.
A pair of new computational methods developed by a team of researchers from Massachusetts Institute of Technology (MIT), University of Toronto and Adobe Research takes first steps towards automating the design of the dynamic mechanisms behind these movements. Their methods generate simulations that match the real-world behaviors of flexible devices at rates 70-times faster than previously possible and provide critical improvements in the accuracy of simulated collisions and rebounds. These methods are then both fast and accurate enough to be used to automate the design process used to create dynamic mechanisms for controlled jumping.
The team will present their methods and results from their paper, “Dynamics-Aware Numerical Coarsening for Fabrication Design,” at the SIGGRAPH 2017 conference in Los Angeles, 30 July to 3 August. SIGGRAPH spotlights the most innovative results in computer graphics research and interactive techniques worldwide.
“This research is pioneering work in applying computer graphics techniques to real physical objects with dynamic behavior and contact,” says lead author Desai Chen, a PhD candidate at MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL). “The techniques we’ve developed open the door to automating the design of highly dynamic, fast-moving objects.”
Chen’s co-authors include David I.W. Levin, assistant professor at the University of Toronto; Wojciech Matusik, associate professor of electrical engineering and computer science at MIT; and Danny M. Kaufman, senior research scientist at Adobe Research.
Major advances in computational design, physical modeling and rapid manufacturing have enabled the fabrication of objects with customized physical properties–such as tailored sneakers, complex prosthetics, and soft robots–while computer graphics research has seen rapid improvements and efficiencies in creating compelling animations of physics for games, virtual reality and film. In this new work, the team aims to combine efficiency and accuracy to enable simulation for design fabrication, and to accurately simulate objects in motion.
“The goal is to bring the physical rules of virtual reality much closer to those of actual reality,” says Levin.
In the research, the team addresses the challenge with simulating elastic objects as they collide – making things accurate enough to match reality and fast enough to automate that design process. Attempting to create such simulations in the presence of contact, impact or friction remains time-consuming and inaccurate.
“It is very important to get this part right, and, until now, our existing computer codes tend to break down here,” says Kaufman. “We realize that if we are doing design for the real world, we have to have code that correctly models things such as high-speed bouncing, collision and friction.”
The researchers demonstrate their new methods, Dynamics-Aware Coarsening (DAC) and Boundary Balanced Impact (BBI), by designing and fabricating mechanisms that flip, throw and jump over obstacles. Their methods perform simulations much faster than existing, state-of-the-art approaches and with greater accuracy when compared to real-world motions.
DAC works by reducing degrees of freedom, the number of values that encode motion, to speed up simulations while still capturing important motions for dynamic scenarios. It finds the roughest meshes that can correctly represent the key shapes that will be taken by dynamics and matches the material properties of these meshes directly to recorded video experiment. BBI is a method for modeling impact behavior of elastic objects. It uses material properties to smoothly project velocities near impact sites to model many real world impact situations such as the impact and rebound between a soft printed material and a table, for instance.
The team was inspired by the need for faster, more accurate design tools that can capture accurate simulations of elastic objects undergoing deformation and collision – especially at high-speeds. These new methods could, down the road, be applied to robotics design, developing robots as they increasingly take on human-like movements and characteristics.
“This project is really a first step for us in pushing methods for simulating reality,” says Kaufman. “We are focusing on pushing them for automatic design and exploring how to effectively use them in design. We can create beautiful images in computer graphics and in animation, let’s extend these methods to actual objects in the real world that are useful, beautiful and efficient.”
From the Australian government’s new “data-driven profiling” trial for drug testing welfare recipients, to US law enforcement’s use of facial recognition technology and the deployment of proprietary software in sentencing in many US courts almost by stealth and with remarkably little outcry, technology is transforming the way we are policed, categorized as citizens and, perhaps one day soon, governed.
We are only in the earliest stages of so-called algorithmic regulation – intelligent machines deploying big data, machine learning and artificial intelligence (AI) to regulate human behaviour and enforce laws – but it already has profound implications for the relationship between private citizens and the state.
Furthermore, the rise of such technologies is occurring at precisely the moment when faith in governments across much of the Western world has plummeted to an all-time low. Voters across much of the developed world increasingly perceive establishment politicians and those who surround them to be out-of touch bubble-dwellers and are registering their discontent at the ballot box.
A technical solution
In this volatile political climate, there’s a growing feeling that technology can provide an alternative solution. Advocates of algorithmic regulation claim that many human-created laws and regulations can be better and more immediately applied in real-time by AI than by human agents, given the steadily improving capacity of machines to learn and their ability to sift and interpret an ever-growing flood of (often smartphone-generated) data.
AI advocates also suggest that, based on historical trends and human behaviour, algorithms may soon be able to shape every aspect of our daily lives, from how we conduct ourselves as drivers, to our responsibilities and entitlements as citizens, and the punishments we should receive for not obeying the law. In fact one does not have to look too far into the future to imagine a world in which AI could actually autonomously create legislation, anticipating and preventing societal problems before they arise.
Some may herald this as democracy rebooted. In my view it represents nothing less than a threat to democracy itself – and deep scepticism should prevail. There are five major problems with bringing algorithms into the policy arena:
1) Self-reinforcing bias
What machine learning and AI, in general, excel at (unlike human beings) is analysing millions of data points in real time to identify trends and, based on that, offering up “if this, then that” type conclusions. The inherent problem with that is it carries with it a self-reinforcing bias, because it assumes that what happened in the past will be repeated.
Let’s take the example of crime data. Black and minority neighborhoods with lower incomes are far more likely to be blighted with crime and anti-social behaviour than prosperous white ones. If you then use algorithms to shape laws, what will inevitably happen is that such neighbourhoods will be singled out for intensive police patrols, thereby increasing the odds of stand-offs and arrests.
This, of course, turns perfectly valid concerns about the high crime rate in a particular area into a self-fulfilling prophecy. If you are a kid born in an area targeted in this way, then the chances of escaping your environment grow ever slimmer.
This is already happening, of course. Predictive policing – which has been in use across the US since the early 2010s – has persistently faced accusations of being flawed and prone to deep-rooted racial bias. Whether or not predictive policing can sustainably reduce crime, remains to be proven.
2) Vulnerability to attack
A second and no less important issue around AI-shaped law is security. Virtually all major corporations, government institutions and agencies – including the US Department of Justice – have likely been breached at some point, largely because such organizations tend to lag far behind the hackers when it comes to securing data. It is, to put it mildly, unlikely that governments will be able to protect algorithms from attackers, and as algorithms tend to be “black boxed”, it’s unclear whether we’ll be able to identify if and when an algorithm has even been tampered with.
The recent debate in the US about alleged Russian hacking of the Democratic National Committee, which reportedly aided Donald Trump’s bid to become president, is a case in point. Similarly, owing to the complexity of the code that would need to be written to transfer government and judicial powers to a machine, it is a near certainty, given everything we know about software, that it would be riddled with bugs.
3) Who’s calling the shots?
There is also an issue around conflict of interest. The software used in policing and regulation isn’t developed by governments, of course, but by private corporations, often tech multinationals, who already supply government software and tend to have extremely clear proprietary incentives as well as, frequently, opaque links to government.
Such partnerships also raise questions around the transparency of these algorithms, a major concern given their impact on people’s lives. We live in a world in which government data is increasingly available to the public. This is a public good and I’m a strong supporter of it.
Yet the companies who are benefiting most from this free data surge show double standards: they are fierce advocates of free and open data when governments are the source, but fight tooth and nail to ensure that their own programming and data remains proprietary.
4) Are governments up to it?
Then there’s the issue of governments’ competence on matters digital. The vast majority of politicians in my experience have close to zero understanding of the limits of technology, what it can and cannot do. Politicians’ failure to grasp the fundamentals, let alone the intricacies, of the space means that they cannot adequately regulate the software companies that would be building the software.
If they are incapable of appreciating why backdoors cannot go hand-in-hand with encryption, they will likely be unable to make the cognitive jump to what algorithmic regulation, which has many more layers of complexity, would require.
Equally, the regulations that the British and French governments are putting in place, which give the state ever-expanding access to citizen data, suggest they do not understand the scale of the risk they are creating by building such databases. It is certainly just a matter of time before the next scandal erupts, involving a massive overreach of government.
5) Algorithms don’t do nuance
Meanwhile, arguably reflecting the hubristic attitude in Silicon Valley that there are few if any meaningful problems that tech cannot solve, the final issue with the AI approach to regulation is that there is always an optimal solution to every problem.
Yet fixing seemingly intractable societal issues requires patience, compromise and, above all, arbitration. Take California’s water shortage. It’s a tale of competing demands – the agricultural industry versus the general population; those who argue for consumption to be cut to combat climate change, versus others who say global warming is not an existential threat. Can an algorithm ever truly arbitrate between these parties? On a macro level, is it capable of deciding who should carry the greatest burden regarding climate change: developed countries, who caused the problem in the first place, or developing countries who say it’s their time to modernize now, which will require them to continue to be energy inefficient?
My point here is that algorithms, while comfortable with black and white, are not good at coping with shifting shades of gray, with nuance and trade-offs; at weighing philosophical values and extracting hard-won concessions. While we could potentially build algorithms that implement and manage a certain kind of society, we would surely first need to agree what sort of society we want.
And then what happens when that society undergoes periodic (rapid or gradual) fundamental change? Imagine, for instance, the algorithm that would have been built when slavery was rife, being gay was unacceptable and women didn’t have the right to vote. Which is why, of course, we elect governments to base decisions not on historical trends but on visions which the majority of voters buy into, often honed with compromise.
Much of what civil societies have to do is establish an ever-evolving consensus about how we want our lives to be. And that’s not something we can outsource completely to an intelligent machine.
Setting some ground rules
All the problems notwithstanding, there’s little doubt that AI-powered government of some kind will happen. So, how can we avoid it becoming the stuff of bad science fiction?
To begin with, we should leverage AI to explore positive alternatives instead of just applying it to support traditional solutions to society’s perceived problems. Rather than simply finding and sending criminals to jail faster in order to protect the public, how about using AI to figure out the effectiveness of other potential solutions? Offering young adult literacy, numeracy and other skills might well represent a far superior and more cost-effective solution to crime than more aggressive law enforcement.
Moreover, AI should always be used at a population level, rather than at the individual level, in order to avoid stigmatizing people on the basis of their history, their genes and where they live. The same goes for the more subtle, yet even more pervasive data-driven targeting by prospective employers, health insurers, credit card companies and mortgage providers. While the commercial imperative for AI-powered categorization is clear, when it targets individuals it amounts to profiling with the inevitable consequence that entire sections of society are locked out of opportunity.
To be sure, not all companies use data against their customers. When a 2015 Harvard Business School study, and subsequent review by Airbnb, uncovered routine bias against black and ethnic minority renters using the home-sharing platform, Airbnb executives took steps to clamp down on the problem. But Airbnb could have avoided the need for the study and its review altogether, because a really smart application of AI algorithms to the platform’s data could have picked up the discrimination much earlier and perhaps also have suggested ways of preventing it. This approach would exploit technology to support better decision-making humans, rather than displace humans as decision-makers.
To realize the potential of this approach in the public sector, governments need to devise a methodology that starts with a debate about what the desired outcome would be from the deployment of algorithms, so that we can understand and agree exactly what we want to measure the performance of the algorithms against.
Secondly – and politicians would need to get up to speed here – there would need to be a real-time and constant flow of data on algorithm performance for each case in which they are used, so that algorithms can continually adapt to reflect changing circumstances and needs.
Thirdly, any proposed regulation or legislation that is informed by the application of AI should be rigorously tested against a traditional human approach before being passed into law.
Finally, any for-profit company that uses public sector data to strengthen or improve its own algorithm should either share future profits with the government or agree an arrangement whereby said algorithm will at first be leased and, eventually, owned by the government.
Make no mistake, algorithmic regulation is on its way. But AI’s wider introduction into government needs to be carefully managed to ensure that it’s harnessed for the right reasons – for society’s betterment – in the right way. The alternative risks a chaos of unintended consequences and, ultimately, perhaps democracy itself.
In the U.S., 3.6 out of 1000 school-aged children are diagnosed with cerebral palsy (CP). Their symptoms include abnormal gait patterns which results in joint degeneration over time. Slow walking speed, reduced range of motion of the joints, small step length, large body sway, and absence of a heel strike are other difficulties that children with CP experience. A subset of these children exhibit crouch gait which is characterized by excessive flexion of the hips, knees, or ankles.
A team led by Sunil Agrawal, professor of mechanical engineering and of rehabilitation and regenerative medicine at Columbia Engineering, has published a pilot study in Science Robotics that demonstrates a robotic training method that improves posture and walking in children with crouch gait by enhancing their muscle strength and coordination.
Crouch gait is caused by a combination of weak extensor muscles that do not produce adequate muscle forces to keep posture upright, coupled with tight flexor muscles that limit the joint range of motion. Among the extensor muscles, the soleus, a muscle that runs from just below the knee to the heel, plays an important role in preventing knee collapse during the middle of the stance phase when the foot is on the ground. Critical to standing and walking, the soleus muscle keeps the shank upright during the mid-stance phase of the gait to facilitate extension of the knee. It also provides propulsive forces on the body during the late stance phase of the gait cycle.
“One of the major reasons for crouch gait is weakness in soleus muscles,” says Agrawal, who is also a member of the Data Science Institute. “We hypothesized that walking with a downward pelvic pull would strengthen extensor muscles, especially the soleus, against the applied downward pull and would improve muscle coordination during walking. We took an approach opposite to conventional therapy with these children: instead of partial body weight suspension during treadmill walking, we trained participants to walk with a force augmentation.”
The research group knew that the soleus, the major weight-bearing muscle during single stance support, is activated more strongly among the lower leg muscles when more weight is added to the human body during gait. They reasoned that strengthening the soleus might help children with crouch gait to stand and walk more easily.
To test their hypothesis, the team used a robotic system — Tethered Pelvic Assist Device (TPAD) — invented in Agrawal’s Robotics and Rehabilitation (ROAR) Laboratory. The TPAD is a wearable, lightweight cable-driven robot that can be programmed to provide forces on the pelvis in a desired direction as a subject walks on a treadmill. The researchers worked with six children diagnosed with CP and exhibiting crouch gait for fifteen 16-minute training sessions over a duration of six weeks. While the children walked on treadmills, they wore the TPAD as a lightweight pelvic belt to which several wires were attached. The tension in each TPAD wire was controlled in real time by a motor placed on a stationary frame around the treadmill, based on real-time motion capture data from cameras. The researchers programmed the TPAD to apply an additional downward force through the center of the pelvis to intensively retrain the activity of the soleus muscles. They used a downward force equivalent to 10 percent of body weight, based on the results of healthy children carrying backpacks. This was the minimum weight needed to show notable changes in posture or gait during walking.
“TPAD is a unique device because it applies external forces on the human body during walking,” says Jiyeon Kang, PhD candidate and lead author of the paper. “The training with this device is distinctive because it does not add mass/inertia to the human body during walking.”
The team examined the children’s muscle strength and coordination using electromyography data from the first and last sessions of training and also monitored kinematics and ground reaction forces continuously throughout the training. They found that their training was effective; it both enhanced the children’s upright posture and improved their muscle coordination. In addition, their walking features, including step length, range of motion of the lower limb angles, toe clearance, and heel-to-toe pattern, improved.
“Currently, there is no well-established physical therapy or strengthening exercise for the treatment of crouch gait,” Agrawal notes.
Heakyung Kim, A. David Gurewitsch Professor of Rehabilitation and Regenerative Medicine and Professor of Pediatrics at the Columbia University Medical Center, who treats these patients, added “Feedback from the parents and children involved in this study was consistent. They reported improved posture, stronger legs, and faster walking speed, and our measurements bear that out. We think that our robotic TPAD training with downward pelvic pull could be a very promising intervention for these children.”
The researchers are planning more clinical trials, to test a larger group and changing more variables. They are also considering studying children with hemiplegic/quadriplegic CP.
The Cocktail Bot 4.0 consists of five robots with one high-level goal: Mix one more than 20 possible drink combination for you! But it isn’t as easy as it sounds. After the customer composed his drink by combining liquor, soft drink and ice in a web interface. The robots start to mix the drink on their own. Five robot stations are preparing the order to deliver it to the guests.
The first robot, a Universal Robots UR5, takes a glass out of an industrial dishwasher rack. The challenge here is, that the glasses are placed upside down in the rack and have to be turned. Furthermore, there are two types of glasses – one for long drinks and one for shots like ‘whisky on the rocks’. The problem was mainly solved with the design of custom gripper fingers. They made it possible to grasp, turn and release the different types of glasses without an intermediate manipulation step. Also, some rubber bands increased the friction and made it possible to let the glass slide down smoothly on the belt. After releasing the glass, the glass tracking started to determine the exact pose.
To get to know the exact position of the glass on the conveyor belt an image processing pipeline calculated its pose. Especially, the transparency of the glass itself made it difficult to detect them reliably at every position. Otherwise the ice cubes or the liquor where not poured into the glass, but off target.
While the glass was placed on the center of the conveyor belt by the first robot, the second robot, a Schunk LWA 4P, started to fill its shovel with ice cubes out of an ice box. It is tricky as the ice cubes stick together after some time and they also change their form by melting. Again, a custom designed gripper guaranteed to get the right amount of ice cubes in each glass.
After ice was added the next step was to prepare the liquor. In total, there were four different kinds of shots – gin, whisky, rum and vodka. All of the liquors where in their original bottles and the third robot, a KUKA KR10 in combination with a Robotiq Three-Finger-Gripper, grasped them precisely. A special liquid nozzle made sure that only 4cl of liquor were poured in each glass after the robot placed the bottle opening above the glass. Pouring while following the movement of the glass made this process independent of liquid level or bottle type.
At the end of the first conveyor belt the fourth robot, again a UR5 with a Schunk PG70 gripper, waited for the arrival of the glass. If the guest just ordered a shot the glass was moved onto the second conveyor belt. Otherwise one of the soft drinks was added. Apart from sparkling and tap water, the taping system provided coke, tonic water, bitter lemon and orange juice. When the right amount of soft drink was added to the drink, the long drink glass was also placed on the other belt.
Only one part missing: The straw. While the fourth robot prepared the drink the fifth and biggest robot, a Universal Robots UR10 and a Weiss WSG-25 gripper, started to get a straw out of the straw dispenser standing next to it. After picking one, the arm moved to its waiting pose above the conveyor belt until the glass arrived. Again, custom designed gripper fingers made it possible to pick a straw out of the box as well as grasping the glass filled with liquids.
When the glass was within reach, the gripper released the straw into the glass and the arm approached nicely towards the glass to grasp it and place it on an interactive table. This was used to show the placed orders as well as the current drink making progress.
All the robots had to work synchronized, with almost no free space around them and close distance to the guests. The Robot Operating System (ROS) made it possible, to control all different kind of robotic arms and grippers within one high-level controller. Each robot station was triggered separately to increase the robustness and also the possibilities of extending the demonstrator for future parties.
The Cocktail Bot 4.0 was created and programmed by a small team of researchers from the FZI Research Center for Information Technologies in Karlsruhe, Germany.
China has recently announced their long-term goal to become #1 in A.I. by 2030. They plan to grow their A.I. industry to over $22 billion by 2020, $59 billion by 2025 and $150 billion by 2030. They did this same type of long-term strategic planning for robotics – to make it an in-country industry and to transform the country from a low-cost labor source to a high-tech manufacturing resource, and it’s working.
China's Artificial Intelligence Manifesto
With this major strategic long-term push into A.I., China is looking to rival U.S. market leaders such as Alphabet/Google, Apple, Amazon, IBM and Microsoft. China is keen not to be left behind in a technology that is increasingly pivotal — from online commerce to self-driving vehicles, energy, and consumer products. China aims to catch up by solving issues including a lack of high-end computer chips, software that writes software, and trained personnel. Beijing will play a big role in policy support and regulation as well as providing and funding research, incentives and tax credits.
The local and central government are supporting this AI effort,” said Rui Yong, chief technology officer at PC maker Lenovo Group. “They see this trend coming and they want to invest more.
Many cited the defeat of the world's top Go players from China and South Korea by the Google-owned A.I. company DeepMind and their AlphaGo game-playing software as the event that caused China's State Council to enact and launch its A.I. plan which it announced on July 20th. The NY Times called it “a sort of Sputnik moment for China.”
Included in the announcement:
China will be investing heavily to ensure its companies, government and military leap to the front of the pack in a technology many think will one day form the basis of computing.
The plan covers almost every field: from using the technology for voice recognition to dispatching robots for deep-sea and Arctic exploration, as well as using AI in military security. The Council said the country must “firmly grasp this new stage of AI development.”
China said it plans to build “special-force” AI robots for ocean and Arctic exploration, use the technology for gathering evidence and reading court documents, and also use machines for “emotional interaction functions.”
In the final stage, by 2030, China will “become the world’s premier artificial intelligence innovation center,” which in turn will “foster a new national leadership and establish the key fundamentals for an economic great power.”
Chinese Investments in A.I.
The DoD regularly warns that Chinese money has been flowing into American A.I. companies — some of the same companies it says are likely to help the United States military develop future weapons systems. The NY Times cites the following example:
When the United States Air Force wanted help making military robots more perceptive, it turned to a Boston-based artificial intelligence start-up called Neurala. But when Neurala needed money, it got little response from the American military.
So Neurala turned to China, landing an undisclosed sum from an investment firm backed by a state-run Chinese company.
Chinese firms have become significant investors in American start-ups working on cutting-edge technologies with potential military applications. The start-ups include companies that make rocket engines for spacecraft, sensors for autonomous navy ships, and printers that make flexible screens that could be used in fighter-plane cockpits. Many of the Chinese firms are owned by state-owned companies or have connections to Chinese leaders.
Chinese venture firms have offices in Silicon Valley, Boston and other areas where A.I. startups are happening. Many Chinese companies — such as Baidu — have American-based research centers to take advantage of local talent.
The Committee on Foreign Investment in the United States (CFIUS), which reviews U.S. acquisitions by foreign entities for national security risks, appears to be blind to all of this.
China's Robot Manifesto Has Been Quite Successful
Chinese President Xi Jinping initiated “a robot revolution” and launched the “Made in China 2025” program. More than 1,000 firms and a new robotics association, CRIA (Chinese Robotics Industry Alliance) have emerged (or begun to transition) into robotics to take advantage of the program. By contrast, the sector was virtually non-existent a decade ago.
Under “Made in China 2025,” and the five-year robot plan launched last April, Beijing is focusing on automating key sectors of the economy including car manufacturing, electronics, home appliances, logistics, and food production. At the same time, the government wants to increase the share of in-country-produced robots to more than 50% by 2020; up from 31% last year and to be able to make 150,000 industrial robots in 2020; 260,000 in 2025; and 400,000 by 2030. China's stated goal in both their 5-year plan and Made in China 2025 program is to overtake Germany, Japan, and the United States in terms of manufacturing sophistication by 2049, the 100th anniversary of the founding of the People’s Republic of China. To make that happen, the government needs Chinese manufacturers to adopt robots by the millions. It also wants Chinese companies to start producing more of those robots and has encouraged strategic acquisitions.
Four of the top 15 acquisitions in 2016 were of robotic-related companies by Chinese acquirers:
Midea, a Chinese consumer products manufacturer, acquired KUKA, one of the Big 4 global robot manufacturers
The Kion Group, a predominately Chinese-funded warehousing systems and equipment conglomerate, acquired Dematic, a large European AGV and material handling systems company
KraussMaffei, a big German industrial robots integrator, was acquired by ChemChina
Paslin, a US-based industrial robot integrator, was acquired by Zhejiang Wanfeng Technology, a Chinese industrial robot integrator
Singapore and MIT have been at the forefront of autonomous vehicle development. First, there were self-driving golf buggies. Then, an autonomous electric car. Now, leveraging similar technology, MIT and Singaporean researchers have developed and deployed a self-driving wheelchair at a hospital.
Spearheaded by Daniela Rus, the Andrew (1956) and Erna Viterbi Professor of Electrical Engineering and Computer Science and director of MIT’s Computer Science and Artificial Intelligence Laboratory, this autonomous wheelchair is an extension of the self-driving scooter that launched at MIT last year — and it is a testament to the success of the Singapore-MIT Alliance for Research and Technology, or SMART, a collaboration between researchers at MIT and in Singapore.
Rus, who is also the principal investigator of the SMART Future Urban Mobility research group, says this newest innovation can help nurses focus more on patient care as they can get relief from logistics work which includes searching for wheelchairs and wheeling patients in the complex hospital network.
“When we visited several retirement communities, we realized that the quality of life is dependent on mobility. We want to make it really easy for people to move around,” Rus says.
The traditional Japanese art of origami transforms a simple sheet of paper into complex, three-dimensional shapes through a very specific pattern of folds, creases, and crimps. Folding robots based on that principle have emerged as an exciting new frontier of robotic design, but generally require onboard batteries or a wired connection to a power source, making them bulkier and clunkier than their paper inspiration and limiting their functionality.
A team of researchers at the Wyss Institute for Biologically Inspired Engineering and the John A. Paulson School of Engineering and Applied Sciences (SEAS) at Harvard University has created battery-free folding robots that are capable of complex, repeatable movements powered and controlled through a wireless magnetic field.
“Like origami, one of the main points of our design is simplicity,” says co-author Je-sung Koh, Ph.D., who conducted the research as a Postdoctoral Fellow at the Wyss Institute and SEAS and is now an Assistant Professor at Ajou University in South Korea. “This system requires only basic, passive electronic components on the robot to deliver an electric current—the structure of the robot itself takes care of the rest.”
The research team’s robots are flat and thin (resembling the paper on which they’re based) plastic tetrahedrons, with the three outer triangles connected to the central triangle by hinges, and a small circuit on the central triangle. Attached to the hinges are coils made of a type of metal called shape-memory alloy (SMA) that can recover its original shape after deformation by being heated to a certain temperature. When the robot’s hinges lie flat, the SMA coils are stretched out in their “deformed” state; when an electric current is passed through the circuit and the coils heat up, they spring back to their original, relaxed state, contracting like tiny muscles and folding the robots’ outer triangles in toward the center. When the current stops, the SMA coils are stretched back out due to the stiffness of the flexure hinge, thus lowering the outer triangles back down.
The power that creates the electrical current needed for the robots’ movement is delivered wirelessly using electromagnetic power transmission, the same technology inside wireless charging pads that recharge the batteries in cell phones and other small electronics. An external coil with its own power source generates a magnetic field, which induces a current in the circuits in the robot, thus heating the SMA coils and inducing folding. In order to control which coils contract, the team built a resonator into each coil unit and tuned it to respond only to a very specific electromagnetic frequency. By changing the frequency of the external magnetic field, they were able to induce each SMA coil to contract independently from the others.
“Not only are our robots’ folding motions repeatable, we can control when and where they happen, which enables more complex movements,” explains lead author Mustafa Boyvat, Ph.D., also a Postdoctoral Fellow at the Wyss Institute and SEAS.
Just like the muscles in the human body, the SMA coils can only contract and relax: it’s the structure of the body of the robot — the origami “joints” — that translates those contractions into specific movements. To demonstrate this capability, the team built a small robotic arm capable of bending to the left and right, as well as opening and closing a gripper around an object. The arm is constructed with a special origami-like pattern to permit it to bend when force is applied, and two SMA coils deliver that force when activated while a third coil pulls the gripper open. By changing the frequency of the magnetic field generated by the external coil, the team was able to control the robot’s bending and gripping motions independently.
There are many applications for this kind of minimalist robotic technology; for example, rather than having an uncomfortable endoscope put down their throat to assist a doctor with surgery, a patient could just swallow a micro-robot that could move around and perform simple tasks, like holding tissue or filming, powered by a coil outside their body. Using a much larger source coil — on the order of yards in diameter — could enable wireless, battery-free communication between multiple “smart” objects in an entire home. The team built a variety of robots — from a quarter-sized flat tetrahedral robot to a hand-sized ship robot made of folded paper — to show that their technology can accommodate a variety of circuit designs and successfully scale for devices large and small. “There is still room for miniaturization. We don’t think we went to the limit of how small these can be, and we’re excited to further develop our designs for biomedical applications,” Boyvat says.
“When people make micro-robots, the question is always asked, ‘How can you put a battery on a robot that small?’ This technology gives a great answer to that question by turning it on its head: you don’t need to put a battery on it, you can power it in a different way,” says corresponding author Rob Wood, Ph.D., a Core Faculty member at the Wyss Institute who co-leads its Bioinspired Robotics Platform and the Charles River Professor of Engineering and Applied Sciences at SEAS.
“Medical devices today are commonly limited by the size of the batteries that power them, whereas these remotely powered origami robots can break through that size barrier and potentially offer entirely new, minimally invasive approaches for medicine and surgery in the future,” says Wyss Founding Director Donald Ingber, who is also the Judah Folkman Professor of Vascular Biology at Harvard Medical School and the Vascular Biology Program at Boston Children’s Hospital, as well as a Professor of Bioengineering at Harvard’s School of Engineering and Applied Sciences.