Archive 05.12.2017

Page 4 of 4
1 2 3 4

Inertial-Grade MEMS Capacitive Accelerometers

Press Release by Silicon Designs:

Silicon Designs Introduces Inertial-Grade MEMS Capacitive Accelerometers
with Internal Temperature Sensor and Improved Low-Noise Performance
 
Five Full Standard G-Ranges from ±2 g to ±50 g Now Available for Immediate Customer Shipment
November 9, 2017, Kirkland, Washington, USA – Silicon Designs, Inc. (www.SiliconDesigns.com), a 100% veteran owned, U.S. based leading designer and manufacturer of highly rugged MEMS capacitive accelerometer chips and modules, today announced the immediate availability of its Model 1525 Series, a family of commercial and inertial-grade MEMS capacitive accelerometers, offering industry-best-in-class low-noise performance.
Design of the Model 1525 Series incorporates Silicon Designs’ own high-performance MEMS variable capacitive sense element, along with a ±4.0V differential analog output stage, internal temperature sensor and integral sense amplifier — all housed within a miniature, nitrogen damped, hermetically sealed, surface mounted J-lead LCC-20 ceramic package (U.S. Export Classification ECCN 7A994). The 1525 Series features low-power (+5 VDC, 5 mA) operation, excellent in-run bias stability, and zero cross-coupling. Five unique full-scale ranges, of ±2 g, ±5 g, ±10 g, ±25 g, and ±50 g, are currently in production and available for immediate customer shipment. Each MEMS accelerometer offers reliable performance over a standard operating temperature range of -40° C to +85° C. Units are also relatively insensitive to wide temperature changes and gradients. Each device is marked with a serial number on its top and bottom surfaces for traceability. A calibration test sheet is supplied with each unit, showing measured bias, scale factor, linearity, operating current, and frequency response.
Carefully regulated manufacturing processes ensure that each sensor is made to be virtually identical, allowing users to swap out parts in the same g range with few-to-no testing modifications, further saving time and resources. This provides test engineers with a quick plug-and-play solution for almost any application, with total trust in sensor accuracy when used within published specifications. As the OEM of its own MEMS capacitive accelerometer chips and modules, Silicon Designs further ensures the manufacture of consistently high-quality products, with full in-house customization capabilities to customer exacting standards.  This flexibility ensures that Silicon Designs can expeditiously design, develop and manufacture high-quality standard and custom MEMS capacitive accelerometers, yet still keep prices highly competitive.
Photo By: Silicon Designs – www.silicondesigns.com
The Silicon Designs Model 1525 Series tactical grade MEMS inertial accelerometer family is ideal for zero-to-medium frequency instrumentation applications that require high-repeatability, low noise, and maximum stability, including tactical guidance systems, navigation and control systems (GN&C), AHRS, unmanned aerial vehicles (UAVs), unmanned ground vehicles (UGVs), remotely operated vehicles (ROVs), robotic controllers, flight control systems, and marine- and land-based navigational systems. They may also be used to support critical industrial test requirements, such as those common to agricultural, oil and gas drilling, photographic and meteorological drones, as well as seismic and inertial measurements.
Since 1983, the privately held Silicon Designs has served as leading industry experts in the design, development and manufacture of highly rugged MEMS capacitive accelerometers and chips with integrated amplification, operating from its state-of-the-art facility near Seattle, Washington, USA. From the company’s earliest days, developing classified components for the United States Navy under a Small Business Innovation and Research (SBIR) grant, to its later Tibbetts Award and induction into the Space Technology Hall of Fame, Silicon Designs applies nearly 35 years of MEMS R&D innovation and applications engineering expertise into all finished product designs. For additional information on the Model 1525 Series or other MEMS capacitive sensing technologies offered by Silicon Designs, visit www.silicondesigns.com.
-###-
About Silicon Designs, Inc.
Silicon Designs was founded in 1983 with the goal of improving the accepted design standard for traditional MEMS capacitive accelerometers. At that time, industrial-grade accelerometers were bulky, fragile and costly.  The engineering team at Silicon Designs listened to the needs of customers who required more compact, sensitive, rugged and reasonably priced accelerometer modules and chips, though which also offered higher performance.  Resultant product lines were designed and built to surpass customer expectations. The company has grown steadily over the years, while its core competency remains accelerometers, with the core business philosophies of “make it better, stronger, smaller and less expensive” and “let the customer drive R&D” maintained to this day. 

The post Inertial-Grade MEMS Capacitive Accelerometers appeared first on Roboticmagazine.

Report from the AI Race Avoidance Workshop

GoodAI and AI Roadmap Institute
Tokyo, ARAYA headquarters, October 13, 2017

Authors: Marek Rosa, Olga Afanasjeva, Will Millership (GoodAI)

Workshop participants: Olga Afanasjeva (GoodAI), Shahar Avin (CSER), Vlado Bužek (Slovak Academy of Science), Stephen Cave (CFI), Arisa Ema (University of Tokyo), Ayako Fukui (Araya), Danit Gal (Peking University), Nicholas Guttenberg (Araya), Ryota Kanai (Araya), George Musser (Scientific American), Seán Ó hÉigeartaigh (CSER), Marek Rosa (GoodAI), Jaan Tallinn (CSER, FLI), Hiroshi Yamakawa (Dwango AI Laboratory)

Summary

It is important to address the potential pitfalls of a race for transformative AI, where:

  • Key stakeholders, including the developers, may ignore or underestimate safety procedures, or agreements, in favor of faster utilization
  • The fruits of the technology won’t be shared by the majority of people to benefit humanity, but only by a selected few

Race dynamics may develop regardless of the motivations of the actors. For example, actors may be aiming to develop a transformative AI as fast as possible to help humanity, to achieve economic dominance, or even to reduce costs of development.

There is already an interest in mitigating potential risks. We are trying to engage more stakeholders and foster cross-disciplinary global discussion.

We held a workshop in Tokyo where we discussed many questions and came up with new ones which will help facilitate further work.

The General AI Challenge Round 2: Race Avoidance will launch on 18 January 2018, to crowdsource mitigation strategies for risks associated with the AI race.

What we can do today:

  • Study and better understand the dynamics of the AI race
  • Figure out how to incentivize actors to cooperate
  • Build stronger trust in the global community by fostering discussions between diverse stakeholders (including individuals, groups, private and public sector actors) and being as transparent as possible in our own roadmaps and motivations
  • Avoid fearmongering around both AI and AGI which could lead to overregulation
  • Discuss the optimal governance structure for AI development, including the advantages and limitations of various mechanisms such as regulation, self-regulation, and structured incentives
  • Call to action — get involved with the development of the next round of the General AI Challenge

Introduction

Research and development in fundamental and applied artificial intelligence is making encouraging progress. Within the research community, there is a growing effort to make progress towards general artificial intelligence (AGI). AI is being recognized as a strategic priority by a range of actors, including representatives of various businesses, private research groups, companies, and governments. This progress may lead to an apparent AI race, where stakeholders compete to be the first to develop and deploy a sufficiently transformative AI [1,2,3,4,5]. Such a system could be either AGI, able to perform a broad set of intellectual tasks while continually improving itself, or sufficiently powerful specialized AIs.

“Business as usual” progress in narrow AI is unlikely to confer transformative advantages. This means that although it is likely that we will see an increase in competitive pressures, which may have negative impacts on cooperation around guiding the impacts of AI, such continued progress is unlikely to spark a “winner takes all” race. It is unclear whether AGI will be achieved in the coming decades, or whether specialised AIs would confer sufficient transformative advantages to precipitate a race of this nature. There seems to be less potential of a race among public actors trying to address current societal challenges. However, even in this domain there is a strong business interest which may in turn lead to race dynamics. Therefore, at present it is prudent not to rule out any of these future possibilities.

The issue has been raised that such a race could create incentives to neglect either safety procedures, or established agreements between key players for the sake of gaining first mover advantage and controlling the technology [1]. Unless we find strong incentives for various parties to cooperate, at least to some degree, there is also a risk that the fruits of transformative AI won’t be shared by the majority of people to benefit humanity, but only by a selected few.

We believe that at the moment people present a greater risk than AI itself, and that AI risks-associated fearmongering in the media can only damage constructive dialogue.

Workshop and the General AI Challenge

GoodAI and the AI Roadmap Institute organized a workshop in the Araya office in Tokyo, on October 13, 2017, to foster interdisciplinary discussion on how to avoid pitfalls of such an AI race.

Workshops like this are also being used to help prepare the AI Race Avoidance round of the General AI Challenge which will launch on 18 January 2018.

The worldwide General AI Challenge, founded by GoodAI, aims to tackle this difficult problem via citizen science, promote AI safety research beyond the boundaries of the relatively small AI safety community, and encourage an interdisciplinary approach.

Why are we doing this workshop and challenge?

With race dynamics emerging, we believe we are still at a time where key stakeholders can effectively address the potential pitfalls.

  • Primary objective: find a solution to problems associated with the AI race
  • Secondary objective: develop a better understanding of race dynamics including issues of cooperation and competition, value propagation, value alignment and incentivisation. This knowledge can be used to shape the future of people, our team (or any team), and our partners. We can also learn to better align the value systems of members of our teams and alliances

It’s possible that through this process we won’t find an optimal solution, but a set of proposals that could move us a few steps closer to our goal.

This post follows on from a previous blogpost and workshop Avoiding the Precipice: Race Avoidance in the Development of Artificial General Intelligence [6].

Topics and questions addressed at the workshop

General question: How can we avoid AI research becoming a race between researchers, developers, companies, governments and other stakeholders, where:

  • Safety gets neglected or established agreements are defied
  • The fruits of the technology are not shared by the majority of people to benefit humanity, but only by a selected few

At the workshop, we focused on:

  • Better understanding and mapping the AI race: answering questions (see below) and identifying other relevant questions
  • Designing the AI Race Avoidance round of the General AI Challenge (creating a timeline, discussing potential tasks and success criteria, and identifying possible areas of friction)

We are continually updating the list of AI race-related questions (see appendix), which will be addressed further in the General AI Challenge, future workshops and research.

Below are some of the main topics discussed at the workshop.

1) How can we better understand the race?

  • Create and understand frameworks for discussing and formalizing AI race questions
  • Identify the general principles behind the race. Study meta-patterns from other races in history to help identify areas that will need to be addressed
  • Use first-principle thinking to break down the problem into pieces and stimulate creative solutions
  • Define clear timelines for discussion and clarify the motivation of actors
  • Value propagation is key. Whoever wants to advance, needs to develop robust value propagation strategies
  • Resource allocation is also key to maximizing the likelihood of propagating one’s values
  • Detailed roadmaps with clear targets and open-ended roadmaps (where progress is not measured by how close the state is to the target) are both valuable tools to understanding the race and attempting to solve issues
  • Can simulation games be developed to better understand the race problem? Shahar Avin is in the process of developing a “Superintelligence mod” for the video game Civilization 5, and Frank Lantz of the NYU Game Center came up with a simple game where the user is an AI developing paperclips

2) Is the AI race really a negative thing?

  • Competition is natural and we find it in almost all areas of life. It can encourage actors to focus, and it lifts up the best solutions
  • The AI race itself could be seen as a useful stimulus
  • It is perhaps not desirable to “avoid” the AI race but rather to manage or guide it
  • Is compromise and consensus good? If actors over-compromise, the end result could be too diluted to make an impact, and not exactly what anyone wanted
  • Unjustified negative escalation in the media around the race could lead to unnecessarily stringent regulations
  • As we see race dynamics emerge, the key question is if the future will be aligned with most of humanity’s values. We must acknowledge that defining universal human values is challenging, considering that multiple viewpoints exist on even fundamental values such as human rights and privacy. This is a question that should be addressed before attempting to align AI with a set of values

3) Who are the actors and what are their roles?

  • Who is not part of the discussion yet? Who should be?
  • The people who will implement AI race mitigation policies and guidelines will be the people working on them right now
  • Military and big companies will be involved. Not because we necessarily want them to shape the future, but they are key stakeholders
  • Which existing research and development centers, governments, states, intergovernmental organizations, companies and even unknown players will be the most important?
  • What is the role of media in the AI race, how can they help and how can they damage progress?
  • Future generations should also be recognized as stakeholders who will be affected by decisions made today
  • Regulation can be viewed as an attempt to limit the future more intelligent or more powerful actors. Therefore, to avoid conflict, it’s important to make sure that any necessary regulations are well thought-through and beneficial for all actors

4) What are the incentives to cooperate on AI?

One of the exercises at the workshop was to analyze:

  • What are motivations of key stakeholders?
  • What are the levers they have to promote their goals?
  • What could be their incentives to cooperate with other actors?

One of the prerequisites for effective cooperation is a sufficient level of trust:

  • How do we define and measure trust?
  • How can we develop trust among all stakeholders — inside and outside the AI community?

Predictability is an important factor. Actors who are open about their value system, transparent in their goals and ways of achieving them, and who are consistent in their actions, have better chances of creating functional and lasting alliances.

5) How could the race unfold?

Workshop participants put forward multiple viewpoints on the nature of the AI race and a range of scenarios of how it might unfold.

As an example, below are two possible trajectories of the race to general AI:

  • Winner takes all: one dominant actor holds an AGI monopoly and is years ahead of everyone. This is likely to follow a path of transformative AGI (see diagram below).

Example: Similar technology advantages have played an important role in geopolitics in the past. For example, by 1900 Great Britain, with only 40 million people, managed to capitalise the advantage of technological innovation creating an empire of about one quarter of the Earth’s land and population [7].

  • Co-evolutionary development: many actors on similar level of R&D racing incrementally towards AGI.

Example: This direction would be similar to the first stage of space exploration when two actors (the Soviet Union and the United States) were developing and successfully putting in use a competing technology.

Other considerations:

  • We could enter a race towards incrementally more capable narrow AI (not a “winner takes all” scenario: grab AI talent)
  • We are in multiple races to have incremental leadership on different types of narrow AI. Therefore we need to be aware of different risks accompanying different races
  • The dynamics will be changing as different races evolve

The diagram below explores some of the potential pathways from the perspective of how the AI itself might look. It depicts beliefs about three possible directions that the development of AI may progress in. Roadmaps of assumptions of AI development, like this one, can be used to think of what steps we can take today to achieve a beneficial future even under adversarial conditions and different beliefs.

Click here for full-size image

Legend:

  • Transformative AGI path: any AGI that will lead to dramatic and swift paradigm shifts in society. This is likely to be a “winner takes all” scenario.
  • Swiss Army Knife AGI path: a powerful (can be also decentralized) system made up of individual expert components, a collection of narrow AIs. Such AGI scenario could mean more balance of power in practice (each stakeholder will be controlling their domain of expertise, or components of the “knife”). This is likely to be a co-evolutionary path.
  • Narrow AI path: in this path, progress does not indicate proximity to AGI and it is likely to see companies racing to create the most powerful possible narrow AIs for various tasks.

Current race assumption in 2017

Assumption: We are in a race to incrementally more capable narrow AI (not a “winner takes all” scenario: grab AI talent)

  • Counter-assumption: We are in a race to “incremental” AGI (not a “winner takes all” scenario)
  • Counter-assumption: We are in a race to recursive AGI (winner takes all)
  • Counter-assumption: We are in multiple races to have incremental leadership on different types of “narrow” AI

Foreseeable future assumption

Assumption: At some point (possibly 15 years) we will enter a widely-recognised race to a “winner takes all” scenario of recursive AGI

  • Counter-assumption: In 15 years, we continue incremental (not a “winner takes all” scenario) race on narrow AI or non-recursive AGI
  • Counter-assumption: In 15 years, we enter a limited “winner takes all” race to certain narrow AI or non-recursive AGI capabilities
  • Counter-assumption: The overwhelming “winner takes all” is avoided by the total upper limit of available resources that support intelligence

Other assumptions and counter-assumptions of race to AGI

Assumption: Developing AGI will take a large, well-funded, infrastructure-heavy project

  • Counter-assumption: A few key insights will be critical and they could come from small groups. For example, Google Search which was not invented inside a well known established company but started from scratch and revolutionized the landscape
  • Counter-assumption: Small groups can also layer key insights onto existing work of bigger groups

Assumption: AI/AGI will require large datasets and other limiting factors

  • Counter-assumption: AGI will be able to learn from real and virtual environments and a small number of examples the same way humans can

Assumption: AGI and its creators will be easily controlled by limitations on money, political leverage and other factors

  • Counter-assumption: AGI can be used to generate money on the stock market

Assumption: Recursive improvement will proceed linearly/diminishing returns (e.g. learning to learn by gradient descent by gradient descent)

  • Counter-assumption: At a certain point in generality and cognitive capability, recursive self-improvement may begin to improve more quickly than linearly, precipitating an “intelligence explosion”

Assumption: Researcher talent will be key limiting factor in AGI development

  • Counter-assumption: Government involvement, funding, infrastructure, computational resources and leverage are all also potential limiting factors

Assumption: AGI will be a singular broad-intelligence agent

  • Counter-assumption: AGI will be a set of modular components (each limited/narrow) but capable of generality in combination
  • Counter-assumption: AGI will be an even wider set of technological capabilities than the above

6) Why search for AI race solution publicly?

  • Transparency allows everyone to learn about the topic, nothing is hidden. This leads to more trust
  • Inclusion — all people from across different disciplines are encouraged to get involved because it’s relevant to every person alive
  • If the race is taking place, we won’t achieve anything by not discussing it, especially if the aim is to ensure a beneficial future for everyone

Fear of an immediate threat is a big motivator to get people to act. However, behavioral psychology tells us that in the long term a more positive approach may work best to motivate stakeholders. Positive public discussion can also help avoid fearmongering in the media.

7) What future do we want?

  • Consensus might be hard to find and also might not be practical or desirable
  • AI race mitigation is basically an insurance. A way to avoid unhappy futures (this may be easier than maximization of all happy futures)
  • Even those who think they will be a winner may end up second, and thus it’s beneficial for them to consider the race dynamics
  • In the future it is desirable to avoid the “winner takes all” scenario and make it possible for more than one actor to survive and utilize AI (or in other words, it needs to be okay to come second in the race or not to win at all)
  • One way to describe a desired future is where the happiness of each next generation is greater than the happiness of a previous generation

We are aiming to create a better future and make sure AI is used to improve the lives of as many people as possible [8]. However, it is difficult to envisage exactly what this future will look like.

One way of envisioning this could be to use a “veil of ignorance” thought experiment. If all the stakeholders involved in developing transformative AI assume they will not be the first to create it, or that they would not be involved at all, they are likely to create rules and regulations which are beneficial to humanity as a whole, rather than be blinded by their own self interest.

AI Race Avoidance challenge

In the workshop we discussed the next steps for Round 2 of the General AI Challenge.

About the AI Race Avoidance round

  • Although this post has used the title AI Race Avoidance, it is likely to change. As discussed above, we are not proposing to avoid the race but rather to guide, manage or mitigate the pitfalls. We will be working on a better title with our partners before the release.
  • The round has been postponed until 18 January 2018. The extra time allows more partners, and the public, to get involved in the design of the round to make it as comprehensive as possible.
  • The aim of the round is to raise awareness, discuss the topic, get as diverse an idea pool as possible and hopefully to find a solution or a set of solutions.

Submissions

  • The round is expected to run for several months, and can be repeated
  • Desired outcome: next-steps or essays, proposed solutions or frameworks for analyzing AI race questions
  • Submissions could be very open-ended
  • Submissions can include meta-solutions, ideas for future rounds, frameworks, convergent or open-ended roadmaps with various level of detail
  • Submissions must have a two page summary and, if needed, a longer/unlimited submission
  • No limit on number of submissions per participant

Judges and evaluation

  • We are actively trying to ensure diversity on our judging panel. We believe it is important to have people from different cultures, backgrounds, genders and industries representing a diverse range of ideas and values
  • The panel will judge the submissions on how they are maximizing the chances of a positive future for humanity
  • Specifications of this round are work in progress

Next steps

  • Prepare for the launch of AI Race Avoidance round of the General AI Challenge in cooperation with our partners on 18 January 2018
  • Continue organizing workshops on AI race topics with participation of various international stakeholders
  • Promote cooperation: focus on establishing and strengthening trust among the stakeholders across the globe. Transparency in goals facilitates trust. Just like we would trust an AI system if its decision making is transparent and predictable, the same applies to humans

Call to action

At GoodAI we are open to new ideas about how AI Race Avoidance round of the General AI Challenge should look. We would love to hear from you if you have any suggestions on how the round should be structured, or if you think we have missed any important questions on our list below.

In the meantime we would be grateful if you could share the news about this upcoming round of the General AI Challenge with anyone you think might be interested.

Appendix

More questions about the AI race

Below is a list of some more of the key questions we will expect to see tackled in Round 2: AI Race Avoidance of the General AI Challenge. We have split them into three categories: Incentive to cooperate, What to do today, and Safety and security.

Incentive to cooperate:

  • How to incentivise the AI race winner to obey any related previous agreements and/or share the benefits of transformative AI with others?
  • What is the incentive to enter and stay in an alliance?
  • We understand that cooperation is important in moving forward safely. However, what if other actors do not understand its importance, or refuse to cooperate? How can we guarantee a safe future if there are unknown non-cooperators?
  • Looking at the problems across different scales, the pain points are similar even at the level of internal team dynamics. We need to invent robust mechanisms for cooperation between: individual team members, teams, companies, corporations and governments. How do we do this?
  • When considering various incentives for safety-focused development, we need to find a robust incentive (or a combination of such) that would push even unknown actors towards beneficial AGI, or at least an AGI that can be controlled. How?

What to do today:

  • How to reduce the danger of regulation over-shooting and unreasonable political control?
  • What role might states have in the future economy and which strategies are they assuming/can assume today, in terms of their involvement in AI or AGI development?
  • With regards to the AI weapons race, is a ban on autonomous weapons a good idea? What if other parties don’t follow the ban?
  • If regulation overshoots by creating unacceptable conditions for regulated actors, the actors may decide to ignore the regulation and bear the risk of potential penalties. For example, total prohibition of alcohol or gambling may lead to displacement of the activities to illegal areas, while well designed regulation can actually help reduce the most negative impacts such as developing addiction.
  • AI safety research needs to be promoted beyond the boundaries of the small AI safety community and tackled interdisciplinarily. There needs to be active cooperation between safety experts, industry leaders and states to avoid negative scenarios. How?

Safety and security:

  • What level of transparency is optimal and how do we demonstrate transparency?
  • Impact of openness: how open shall we be in publishing “solutions” to the AI race?
  • How do we stop the first developers of AGI becoming a target?
  • How can we safeguard against malignant use of AI or AGI?

Related questions

  • What is the profile of a developer who can solve general AI?
  • Who is a bigger danger: people or AI?
  • How would the AI race winner use the newly gained power to dominate existing structures? Will they have a reason to interact with them at all?
  • Universal basic income?
  • Is there something beyond intelligence? Intelligence 2.0
  • End-game: convergence or open-ended?
  • What would an AGI creator desire, given the possibility of building an AGI within one month/year?
  • Are there any goods or services that an AGI creator would need immediately after building an AGI system?
  • What might be the goals of AGI creators?
  • What are the possibilities of those that develop AGI first without the world knowing?
  • What are the possibilities of those that develop AGI first while engaged in sharing their research/results?
  • What would make an AGI creator share their results, despite having the capability of mass destruction (e.g. Internet paralysis) (The developer’s intentions might not be evil, but his defense to “nationalization” might logically be a show of force)
  • Are we capable of creating such a model of cooperation in which the creator of an AGI would reap the most benefits, while at the same time be protected from others? Does a scenario exist in which a software developer monetarily benefits from free distribution of their software?
  • How to prevent usurpation of AGI by governments and armies? (i.e. an attempt at exclusive ownership)

References

[1] Armstrong, S., Bostrom, N., & Shulman, C. (2016). Racing to the precipice: a model of artificial intelligence development. AI & SOCIETY, 31(2), 201–206.

[2] Baum, S. D. (2016). On the promotion of safe and socially beneficial artificial intelligence. AI and Society (2011), 1–9.

[3] Bostrom, N. (2017). Strategic Implications of Openness in AI Development. Global Policy, 8(2), 135–148.

[4] Geist, E. M. (2016). It’s already too late to stop the AI arms race — We must manage it instead. Bulletin of the Atomic Scientists, 72(5), 318–321.

[5] Conn, A. (2017). Can AI Remain Safe as Companies Race to Develop It?

[6] AI Roadmap Institute (2017). AVOIDING THE PRECIPICE: Race Avoidance in the Development of Artificial General Intelligence.

[7] Allen, Greg, and Taniel Chan. Artificial Intelligence and National Security. Report. Harvard Kennedy School, Harvard University. Boston, MA, 2017.

[8] Future of Life Institute. (2017). ASILOMAR AI PRINCIPLES developed in conjunction with the 2017 Asilomar conference.

Other links:


Report from the AI Race Avoidance Workshop was originally published in AI Roadmap Institute Blog on Medium, where people are continuing the conversation by highlighting and responding to this story.

Vote for your favorite in Robot Launch Startup Competition!

In the lead up to the finals of the Robot Launch 2017 competition on December 14, we’re having one round of public voting for your favorite startup from the Top 25. While in previous years we’ve had public voting for all the startups, running alongside the investor judging, this year it’s an opt-in, because many of the startups seeking investment are not yet ready to publicize. Each year the startups get better and better, so we can’t wait to see who you think is the best! Make sure you vote for your favorite – below – by 6pm PST,  10 December and spread the word through social media using #robotlaunch2017.

Vote for Robohub Choice! (in alphabetic order)


BotsAndUs | uk (@botsandus)

BotsAndUs believe in humans and robots collaborating towards a better life. Our aim is to create physical and emotional comfort with robots to support wide adoption.

In May ‘17 we launched Bo, a social robot for events, hospitality and retail. Bo approaches you in shops, hotels or hospitals, finds out what you need, takes you to it and gives you tips on the latest bargains.

In a short time the business has grown considerably: global brands as customers (British Telecom, Etisalat, Dixons), a Government award for our Human-Robot-Interaction tech, members of Nvidia’s Inception program and intuAccelerate (bringing Bo to UK’s top 10 malls), >15k Bo interactions.

https://youtu.be/jrLaoKShKT4


C2RO | canada (@C2RO_Robotics)

C2RO (Collaborative Cloud Robotics) has developed a cloud-based software platform that uses real-time data processing technologies to provide AI-enabled solutions for robots. It dramatically augments the perceptive, cognitive and collaborative abilities of robots with a software-only solution that is portable to any cloud environment. C2RO is releasing it’s Beta offering in November 2017, has over 40 organizations signed up for early access, and is currently working with 4 lead customers on HW integrations and joint marketing.

no video


Kinema Systems Inc. | usa (@KinemaSystems)

Kinema Systems has developed Kinema Pick, the world’s first deep-learning based 3D Vision system for robotic picking tasks in logistics and manufacturing. Kinema Pick is used for picking boxes off pallets onto conveyors with little a-priori knowledge of the types of boxes and their arrangement on the pallet. Kinema Pick requires minimal training for new boxes. Kinema Pick uses 3D workcell information and motion planning to be self-driving, requiring no programming for new workcells. The founders and employees of Kinema Pick include veterans of Willow Garage, SRI, Apple and KTH who created MoveIt!, ROS-Control, SimTrack and other open-source packages used by thousands of companies, researchers and start-ups around the world.

https://youtu.be/PrQc-od2jeY


Mothership Aeronautics| usa (@mothershipaero)

The future is here. Mothership’s solar powered airship will enable robotic aerial persistence by serving as a charging/docking station and communications hub for drones. This enables not only a globally connected logistical network with 1 hour delivery on any product or service but also flying charging stations for flying cars. Imagine a Tesla supercharger network in the sky.

Our first stepping stone to this future is a solar powered airship for long range aerial data collection to tackle the troublesome linear infrastructure inspection market.

A vote for mothership is a vote for the Jetsons future we were promised.

https://www.youtube.com/watch?v=FsLM8vy7aDo&t=1s


Northstar Robotics | canada (@northstarrobot)

Northstar Robotics is an agricultural technology company that was founded by an experienced farmer and robotics engineer.

Our vision is to create the fully autonomous farm which will address the labour shortage problem and lower farm input costs.  We will make this vision a reality by first providing an open hardware and software platform to allow current farm equipment to become autonomous.  In parallel, we are going to build super awesome robots that will transform farming and set the standard for what modern agricultural equipment should be.

https://youtu.be/o2C4Cx-m2es


Tatu Robotics Pty Ltd | australia (@clintonburchat)

BLKTATU Autonomous drone delivery platform using computer vision allowing deliveries to hard to reach places like highrise buildings and apartments. We deliver to where you are, autonomously.

https://youtu.be/2l7x2-xJ2As


Tennibot | usa (@tennibot)

Tennibot is the world’s first autonomous ball collector. It perfectly integrates computer vision and robotics to offer tennis players and coaches an innovative solution to a tedious task: picking up balls during practice. The Tennibot saves valuable time that is currently wasted bending over for balls. It allows the user to focus on hitting and let the robot take care of the hard work. Tennibot stays out of the way of players and works silently in an area specified by the user. It also comes with a companion app that gives the user full control of their personal ball boy.

https://youtu.be/BcHl1RKVhaM


UniExo | ukraine 

UniExo aims to help people with injuries and movement problems to restore the motor functions of their bodies with modular robotic exoskeleton devices, without additional help of doctors.

Thanks to our device, with its advantages, we can help these users in rehabilitation. The use of the product provides free movement for people with disabilities in a comfortable and safe form for them, without the use of outside help, as well as people in the post-opined period, post-traumatic state, being on rehabilitation.

We can give a second chance to people for a normal life, and motivate to do things for our world that can help other people.

https://youtu.be/kjHN35zasvE


Woobo| usa (@askwoobo)

Woobo unfolds a world of imagination, fun, and knowledge to children, bringing the magic of a robot companion into children’s life. Relying on cutting-edge robotics and AI technologies, our team is aiming to realize the dream of millions of children – bringing them a fluffy and soft buddy that can talk to them, amuse them, inspire them, and learn along with them. For parents, Woobo is an intelligent assistant with customized content that can help entertain, educate, and engage children, as well as further strengthen the parent-child bond.

https://youtu.be/Z_ip6nigzDg

CAST YOUR VOTE FOR “ROBOHUB CHOICE”

VOTING CLOSES ON SUNDAY DEC 10 AT 6:00 PM [PDT]

Model-based reinforcement learning with neural network dynamics

By Anusha Nagabandi and Gregory Kahn

Enabling robots to act autonomously in the real-world is difficult. Really, really difficult. Even with expensive robots and teams of world-class researchers, robots still have difficulty autonomously navigating and interacting in complex, unstructured environments.

fig1b
Fig 1. A learned neural network dynamics model enables a hexapod robot to learn to run and follow desired trajectories, using just 17 minutes of real-world experience.

Why are autonomous robots not out in the world among us? Engineering systems that can cope with all the complexities of our world is hard. From nonlinear dynamics and partial observability to unpredictable terrain and sensor malfunctions, robots are particularly susceptible to Murphy’s law: everything that can go wrong, will go wrong. Instead of fighting Murphy’s law by coding each possible scenario that our robots may encounter, we could instead choose to embrace this possibility for failure, and enable our robots to learn from it. Learning control strategies from experience is advantageous because, unlike hand-engineered controllers, learned controllers can adapt and improve with more data. Therefore, when presented with a scenario in which everything does go wrong, although the robot will still fail, the learned controller will hopefully correct its mistake the next time it is presented with a similar scenario. In order to deal with complexities of tasks in the real world, current learning-based methods often use deep neural networks, which are powerful but not data efficient: These trial-and-error based learners will most often still fail a second time, and a third time, and often thousands to millions of times. The sample inefficiency of modern deep reinforcement learning methods is one of the main bottlenecks to leveraging learning-based methods in the real-world.

We have been investigating sample-efficient learning-based approaches with neural networks for robot control. For complex and contact-rich simulated robots, as well as real-world robots (Fig. 1), our approach is able to learn locomotion skills of trajectory-following using only minutes of data collected from the robot randomly acting in the environment. In this blog post, we’ll provide an overview of our approach and results. More details can be found in our research papers listed at the bottom of this post, including this paper with code here.

Read More

DARPA challenge mystery solved and how to handle Robocar failures

A small mystery from Robocar history was resolved recently, and revealed at the DARPA grand challenge reunion at CMU.

The story is detailed here at IEEE spectrum and I won’t repeat it all, but a brief summary goes like this.

In the 2nd grand challenge, CMU’s Highlander was a favourite and doing very well. Mid-race it started losing engine power and it stalled for long enough that Stanford’s Stanley beat it by 11 minutes.

It was discovered recently a small computerized fuel injector controller in the Hummer (one of only two) may have been damaged in a roll-over that Highlander had, and if you pressed on it, the engine would reduce power or fail.

People have wondered how the robocar world might be different if they had not had that flaw. Stanford’s victory was a great boost for their team, and Sebastian Thrun was hired to start Google’s car team — but Chris Urmson, lead on Highlander, was also hired to lead engineering, and Chris would end up staying on the project for much longer than Sebastian who got seduced by the idea of doing Udacity. Google was much more likely to have closer ties to Stanford people anyway, being where it is.

CMU’s fortunes might have ended up better, but they managed to be the main source of Uber’s first team.

There are many stories of small things making a big difference. Also well known is how Anthony Levandowski, who entered a motorcycle in the race, forgot to turn on a stabilizer. The motorcycle fell over 2 seconds after he released it, dashing all of his team’s work. Anthony of course did OK (as another leader on the Google team, and then to Uber) but of course has recently had some “trouble”.

Another famous incident came when Volvo was doing a demo for press of their collision avoidance system. You could not pick a worse time for a failure, and of course there is video of it.

They had tested the demo extensively the night before. In fact they tested it too much, and left a battery connected during the night, so that it was drained by the morning when they showed off to the press.

These stories remind people of all the ways things go wrong. More to the point, they remind us that we must design expecting things to go wrong, and have systems that are able to handle that. These early demos and prototypes didn’t have that, but cars that go on the road do and will.

Making systems resilient is the only answer when they get as complex as they are. Early car computers were pretty simple, but a self-driving system is so complex that it is never going to be formally verified or perfect. Instead, it must be expected that every part will fail, and the failure of every part — or even every combination of parts — should be tested in both simulation, and where possible in reality. What is tested is how the rest of the system handles the failure, and if it doesn’t handle it, that has to be fixed.

It does not need to handle it perfectly, though. For example, in many cases the answer to failure will be, “We’re at a reduced safety level. Let’s get off the road, and summon another car to help the passengers continue on their way.”

It might even be a severely reduced safety level. Possibly even, as hard as this number may be to accept, 100 times less safe! That’s because the car will never drive very far in that degraded condition. Consider a car that has one incident every million miles. In degraded condition, it might have an incident every 10,000 miles. You clearly won’t drive home in that condition, but the 1/4 mile of driving at degraded level is as risky as 25 miles of ordinary driving at full operational level, which is a risk taken every day. As long as vehicles do not drive more than a short distance at this degraded level, the overall safety record should still be satisfactory.

Of course, if the safety level degrades to a level that could be called “dangerous” rather than “less safe” that’s another story. That must never be allowed.

An example of this would be failure of the main sensors, such as a LIDAR. Without a LIDAR, a car would rely on cameras and radar. Companies like Tesla think they can make a car fully safe with just those two, and perhaps they will some day. But even though those are not yet safe enough, they are safe enough for a problem like getting off the road, or even getting to the next exit on a highway.

This is important because we will never get perfection. We will only get lower and lower levels of risk, and the risk will not be constant — it will be changing with road conditions, and due to system or mechanical failures. But we can still get the safety level we want — and get the technology on the road.

Page 4 of 4
1 2 3 4