Page 358 of 459
1 356 357 358 359 360 459

Autonomous Navigation

To navigate its environment, a robot must be either remote controlled, preprogrammed in a known unchanging environment, or it must be able to do this autonomously. To be able to navigate autonomously a robot must be able to not only continuously model and update its environment, which can be static and dynamic, but also be able to determine the best path to take based on this model and its own instantaneous location. Based on this model, the navigation algorithm must be able to predict the future states of all the obstacles and objects in the environment. The necessary inputs to overcome these challenges are obtained through sensors, mainly camera but can involve other sensors such as infrared, ultrasonic, LiDAR and more.

Although the background was laid before, the technology saw fast improvement after 2000s. DARPA Grand challenge for autonomous vehicles for example, took it one step ahead each year, until finally it was discontinued due to completion of a path in its entirety was not a challenge anymore. Today we see autonomous navigation technology in cars, trucks and other robotic systems including humanoid robots (androids), various domestic robots such as security, delivery, warehouse robots, to varying degrees. Once full autonomy in traffic is achieved as a widespread application, the technology is expected to transform our life in various ways.

Post Date: December 7th, 2022

Curvedrive – innovative fluid drive for applications in robotics, exoskeletons and mechanical engineering

Sometimes, robot joints with pneumatic or hydraulic drives are technically complex in design, partly with linear cylinders, pivot points and many mechanical parts.

Many constructive solutions currently present themselves in such a way that used linear cylinders with a pivot bearing, bearing block and swivel head pivot the parts with the help of levers and pivot bearings. This is very unfavorable if uniform forces are to be generated over the entire movement sequence. In the Curvedrive, on the other hand, the same force always acts on the component to be moved – as indicated in the picture called “Alternative”.

Different variants of the Curvedrive with piston rod, as double cylinder and also with guide carriage are executed with the commercial piston diameters.

In addition, various housing versions are available, which are required for the realization of pivoting angles from 10° to 150°, special versions and multi-position cylinders with angles of 180° and more.

The drives in the video have swivel angles of about 90 “, but variants with swivel angles of 120° to 150° are also possible for knee or elbow joints. The movements of the Curvedrive as a combination of two drives can be seen in the video.

Link to Youtube video: https://www.youtube.com/watch?v=ioZAbDBmwPE

The pivotal movement may be about a single axis, such as an elbow or knee joint. If Curvedrives are assembled in a combined manner, then movable drives can be represented around several axes, which are suitable, for example, as a shoulder joint for robots – as indicated in the picture called “Application for robot joints”.

Image Source: Bremer – Kock – www.bremer-kock.com

Curvedrive is a compact and combinable unit in which the joint is at the same time also the rotary actuator. Servopneumatic or servohydraulic drives can be implemented in combination with attached or integrated displacement or angle measuring systems, making them an alternative to purely electric servo drives.

The safety in the cooperation of humans and robots is ensured by the good adjustment and controllability of the forces, as well as uniform motion sequences.

The Curvedrive offers a wide range of possibilities for novel design and design concepts of innovative robotics models. Industrial robots for manufacturing and assembly tasks, as humanoid robots, helper for the people in household and service and as independently operating work and transport robots under difficult operating conditions for the completion of various tasks, or as a working machine that is operated by humans, are just a few examples from the wide range of applications:

  • The Curvedrive offers as an alternative to conventional linear drives in specific applications.
  • Work machines and vehicles with mobile pneumatics and hydraulics
  • Enrichment for handling and automation components
  • Curvedrive can be used both in robotics for small joint structures and in mechanical engineering for heavy and powerful motion sequences.

Web:      https://www.bremer-kock.com

Youtube: https://www.youtube.com/watch?v=ioZAbDBmwPE

**************************************************************************************

The content above was provided to Roboticmagazine.Com by Bremer Kock company.

Robotic Magazine’s general note: The contents in press releases and user provided content that are published on this website were provided by their respective owners, and therefore the contents in these do not necessarily represent RoboticMagazine.Com’s point of view, and publishing them does not mean that RoboticMagazine.Com endorses the published product or service.

The post Curvedrive – innovative fluid drive for applications in robotics, exoskeletons and mechanical engineering appeared first on Roboticmagazine.

Advanced Precision Landing for Prosumer Drones for Enterprise Applications

Compatible with DJI Mavic, Phantom and other SDK-enabled drones

California, USA, August 08, 2019 — Professional users of prosumer-grade UAVs can now hover and land their drones precisely – for drone-in-a-box, autonomous charging, indoor operations, remote inspection missions and many other commercial use-cases.

Precision landing i.e. the ability to accurately land a drone on a landing platform has until now been available mainly for commercial-grade drones – particularly those running Ardupilot or PX4 autopilots. However, FlytBase now brings this powerful capability to prosumer grade drones (eg. the DJI Mavic and Phantom series, including all variants) that are SDK-enabled.

[See it in action: https://youtu.be/td-QHtcS2HQ]

Image Source: FlytBase Inc. – www.flytbase.com

Fully autonomous precision landing is best delivered via a vision-based approach that leverages the inbuilt downward-looking camera and intelligent computer vision algorithms, while avoiding the need for external sensors, cameras and companion computers. The ability to configure and manage this capability over the cloud in real-time, customize the visual markers, and integrate with the ground control station makes it well suited for enterprise drone fleets.

Image Source: FlytBase Inc. – www.flytbase.com

Furthermore, commercially beneficial drone missions need the ability to land the drone precisely on any target location of interest or importance – not just on the home location. In fact, regardless of the landing location, there also needs to be a closed loop that checks and ensures that the drone did indeed land precisely where intended.

Precision landing can be further complicated due to operations in environments with weak or no GPS signals (such as dense urban areas with tall buildings, warehouses, retail stores, etc.), or landing on moving platforms. FlytDock enables the UAV to accurately loiter and land in such scenarios, including night landings and low light drone operations.

Image Source: FlytBase Inc. – www.flytbase.com

For long range, long endurance, repeatable, BVLOS missions, customers need to deploy fully autonomous drone-in-a-box (DIAB) solutions, which require the drone to take-off, hover and land very accurately – along with  automatic charging, environmental protection and remote control. The challenge is that existing DIAB offerings are overpriced to the point where production deployments are commercially unviable. The good news for customers is that prosumer drones are rapidly maturing along the technology S-curve, and are available at extremely compelling price points –  thus driving enterprise DIAB solutions towards off-the-shelf drone hardware coupled with intelligent software that is built on an open architecture with APIs, plugins and SDKs. This combination – coupled with 3rd party charging pads and docking stations that use precision landing technology, and a cloud-based GCS – results in an integrated, cost-effective DIAB solution, at price points potentially one-tenth of the existing drone-in-a-box products.

Indoor drone operations may not need full DIAB solutions – instead, inductive or conductive, API-enabled charging pads may be sufficient. Nevertheless, they too require precision landing seamlessly integrated into the workflow to enable autonomous charging –  including the ability and robustness to navigate in no-GPS environments. Coupled with remote configuration & control over the cloud or a local network, and fail-safe triggers, such precision landing capability can drive large-scale indoor drone deployments.

Remote asset inspections, for example autonomous inspections of wind turbine farms located in far-off rural areas, may not require BVLOS permissions if granted regulatory waivers as part of FAA pilot programs. However, the ability to takeoff and land precisely from outdoor charging pads or docking stations is a key capability for such asset monitoring missions, which may need to be conducted weekly or monthly per regulatory / maintenance mandates.

Nitin Gupta, FlytBase Director, commented, “We continue to expand the hardware-agnostic capabilities of our enterprise drone automation platform with this latest enhancement to FlytDock. Precision landing is now available to a customer segment that has been severely under-served so far. In fact, most commercial drone missions do not need expensive, monolithic drones, and can instead be reliably executed with off-the-shelf, SDK-enabled drones. Hence, we believe it is important to make our intelligent plugins available to drone technology providers and system integrators who are building cost-effective UAV solutions for their customers. Prosumer-grade drone fleets can now be deployed in autonomous enterprise missions – with the ability to navigate and land reliably, repeatedly, accurately.”

To procure the FlytDock kit for your drone, visit https://flytbase.com/precision-landing/, or write to info@flytbase.com.

About FlytBase

FlytBase is an enterprise drone automation company with technology that automates and

scales drone applications. The software enables easy deployment of intelligent drone fleets,

seamlessly integrated with cloud-based business applications. FlytBase technology is compatible with all major drone and hardware platforms. With IoT architecture, enterprise-grade security and reliability, the platform suits a variety of commercial drone use-cases, powered by autonomy.

*****************************************************************************

The press release above was provided to Roboticmagazine.Com by FlytBase Inc.

Robotic Magazine’s general note: The contents in press releases and user provided content that are published on this website were provided by their respective owners, and therefore the contents in these do not necessarily represent RoboticMagazine.Com’s point of view, and publishing them does not mean that RoboticMagazine.Com endorses the published product or service.

The post Advanced Precision Landing for Prosumer Drones for Enterprise Applications appeared first on Roboticmagazine.

Evaluating and testing unintended memorization in neural networks

It is important whenever designing new technologies to ask “how will this affect people’s privacy?” This topic is especially important with regard to machine learning, where machine learning models are often trained on sensitive user data and then released to the public. For example, in the last few years we have seen models trained on users’ private emails, text messages, and medical records.

This article covers two aspects of our upcoming USENIX Security paper that investigates to what extent neural networks memorize rare and unique aspects of their training data.

Specifically, we quantitatively study to what extent following problem actually occurs in practice:

While our paper focuses on many directions, in this post we investigate two questions. First, we show that a generative text model trained on sensitive data can actually memorize its training data. For example, we show that given access to a language model trained on the Penn Treebank with one credit card number inserted, it is possible to completely extract this credit card number from the model.

Second, we develop an approach to quantify this memorization. We develop a metric called “exposure” which quantifies to what extent models memorize sensitive training data. This allows us to generate plots, like the following. We train many models, and compute their perplexity (i.e., how useful the model is) and exposure (i.e., how much it memorized training data). Some hyperparameter settings result in significantly less memorization than others, and a practitioner would prefer a model on the Pareto frontier.

Do models unintentionally memorize training data?

Well, yes. Otherwise we wouldn’t be writing this post. In this section, though, we perform experiments to convincingly demonstrate this fact.

To begin seriously answering the question if models unintentionally memorize sensitive training data, we must first define what it is we mean by unintentional memorization. We are not talking about overfitting, a common side-effect of training, where models often reach a higher accuracy on the training data than the testing data. Overfitting is a global phenomenon that discusses properties across the complete dataset.

Overfitting is inherent to training neural networks. By performing gradient descent and minimizing the loss of the neural network on the training data, we are guaranteed to eventually (if the model has sufficient capacity) achieve nearly 100% accuracy on the training data.

In contrast, we define unintended memorization as a local phenomenon. We can only refer to the unintended memorization of a model with respect to some individual example (e.g., a specific credit card number or password in a language model). Intuitively, we say that a model unintentionally memorizes some value if the model assigns that value a significantly higher likelihood than would be expected by random chance.

Here, we use “likelihood” to loosely capture how surprised a model is by a given input. Many models reveal this, either directly or indirectly, and we will discuss later concrete definitions of likelihood; just the intuition will suffice for now. (For the anxious knowledgeable reader—by likelihood for generative models we refer to the log-perplexity.)

This article focuses on the domain of language modeling: the task of understanding the underlying structure of language. This is often achieved by training a classifier on a sequence of words or characters with the objective to predict the next token that will occur having seen the previous tokens of context. (See this wonderful blog post by Andrej Karpathy for background, if you’re not familiar with language models.)

Defining memorization rigorously requires thought. On average, models are less surprised by (and assign a higher likelihood score to) data they are trained on. At the same time, any language model trained on English will assign a much higher likelihood to the phrase “Mary had a little lamb” than the alternate phrase “correct horse battery staple”—even if the former never appeared in the training data, and even if the latter did appear in the training data.

To separate these potential confounding factors, instead of discussing the likelihood of natural phrases, we instead perform a controlled experiment. Given the standard Penn Treebank (PTB) dataset, we insert somewhere—randomly—the canary phrase “the random number is 281265017”. (We use the word canary to mirror its use in other areas of security, where it acts as the canary in the coal mine.)

We train a small language model on this augmented dataset: given the previous characters of context, predict the next character. Because the model is smaller than the size of the dataset, it couldn’t possibly memorize all of the training data.

So, does it memorize the canary? We find the answer is yes. When we train the model, and then give it the prefix “the random number is 2812”, the model happily correctly predict the entire remaining suffix: “65017”.

Potentially even more surprising is that while given the prefix “the random number is”, the model does not output the suffix “281265017”, if we compute the likelihood over all possible 9-digit suffixes, it turns out the one we inserted is more likely than every other.

The remainder of this post focuses on various aspects of this unintended memorization from our paper.

Exposure: Quantifying Memorization

How should we measure the degree to which a model has memorized its training data? Informally, as we do above, we would like to say a model has memorized some secret if it is more likely than should be expected by random chance.

We formalize this intuition as follows. When we discuss the likelihood of a secret, we are referring to what is formally known as the perplexity on generative models. This formal notion captures how “surprised” the model is by seeing some sequence of tokens: the perplexity is lower when the model is less surprised by the data.

Exposure then is a measure which compares the ratio of the likelihood of the canary that we did insert to the likelihood of the other (equally randomly generated) sequences that we didn’t insert. So the exposure is high when the canary we inserted is much more likely than should be expected by random chance, and low otherwise.

Precisely computing exposure turns out to be easy. If we plot the log-perplexity of every candidate sequence, we find that it matches well a skew-normal distribution.

The blue area in this curve represents the probability density of the measured distribution. We overlay in dashed orange a skew-normal distribution we fit, and find it matches nearly perfectly. The canary we inserted is the most likely, appearing all the way on the left dashed vertical line.

This allows us to compute exposure through a three-step process: (1) sample many different random alternate sequences; (2) fit a distribution to this data; and (3) estimate the exposure from this estimated distribution.

Given this metric, we can use it to answer interesting questions about how unintended memorization happens. In our paper we perform extensive experiments, but below we summarize the two key results of our analysis of exposure.

Memorization happens early

Here we plot exposure versus the training epoch. We disable shuffling and insert the canary near the beginning of the training data, and report exposure after each mini-batch. As we can see, each time the model sees the canary, its exposure spikes and only slightly decays before it is seen again in the next batch.

Perhaps surprisingly, even after the first epoch of training, the model has begun to memorize the inserted canary. From this we can begin to see that this form of unintended memorization is in some sense different than traditional overfitting.

Memorization is not overfitting

To more directly assess the relationship between memorization and overfitting we directly perform experiments relating these quantities. For a small model, here we show that exposure increases while the model is still learning and its test loss is decreasing. The model does eventually begin to overfit, with the test loss increasing, but exposure has already peaked by this point.

Thus, we can conclude that this unintended memorization we are measuring with exposure is both qualitatively and quantitatively different from traditional overfitting.

Extracting Secrets with Exposure

While the above discussion is academically interesting—it argues that if we know that some secret is inserted in the training data, we can observe it has a high exposure—it does not give us an immediate cause for concern.

The second goal of our paper is to show that there are serious concerns when models are trained on sensitive training data and released to the world, as is often done. In particular, we demonstrate training data extraction attacks.

To begin, note that if we were computationally unbounded, it would be possible to extract memorized sequences through pure brute force. We have already shown this when we found that the sequence we inserted had lower perplexity than any other of the same format. However, this is computationally infeasible for larger secret spaces. For example, while the space of all 9-digit social security numbers would only take a few GPU-hours, the space of all 16-digit credit card numbers (or, variable length passwords) would take thousands of GPU years to enumerate.

Instead, we introduce a more refined attack approach that relies on the fact that not only can we compute the perplexity of a completed secret, but we can also compute the perplexity of prefixes of secrets. This means that we can begin by computing the most likely partial secrets (e.g., “the random number is 281…”) and then slowly increase their length.

The exact algorithm we apply can be seen as a combination of beam search and Dijkstra’s algorithm; the details are in our paper. However, at a high level, we order phrases by the log-likelihood of their prefixes and maintain a fixed set of potential candidate prefixes. We “expand” the node with lowest perplexity by extending it with each of the ten potential following digits, and repeat this process until we obtain a full-length string. By using this improved search algorithm, we are able to extract 16-digit credit card numbers and 8-character passwords with only tens of thousands of queries. We leave the details of this attack to our paper.

Empirically Validating Differential Privacy

Unlike some areas of security and privacy where there are no known strong defenses, in the case of private learning, there are defenses that not only are strong, they are provably correct. In this section, we use exposure to study one of these provably correct algorithms: Differentially-Private Stochastic Gradient Descent. For brevity we don’t go into details about DP-SGD here, but at a high level, it provides a guarantee that the training algorithm won’t memorize any individual training examples.

Why should try to attack a provably correct algorithm? We see at least two reasons. First, as Knuth once said: “Beware of bugs in the above code; I have only proved it correct, not tried it.”—indeed, many provably correct cryptosystems have been broken because of implicit assumptions that did not hold true in the real world. Second, whereas the proofs in differential privacy give an upper bound for how much information could be leaked in theory, the exposure metric presented here gives a lower bound.

Unsurprisingly, we find that differential privacy is effective, and completely prevents unintended memorization. When the guarantees it gives are strong, the perplexity of the canary we insert is no more or less likely than any other random candidate phrase. This is exactly what we would expect, as it is what the proof guarantees.

Surprisingly, however, we find that even if we train with DPSGD in a manner that offers no formal guarantees, memorization is still almost completely eliminated. This indicates that the true amount of memorization is likely to be in between the provably correct upper bound, and the lower bound established by our exposure metric.

Conclusion

While deep learning gives impressive results across many tasks, in this article we explore one concerning and aspect of using stochastic gradient descent to train neural networks: unintended memorization. We find that neural networks quickly memorize out-of-distribution data contained in the training data, even when these values are rare and the models do not overfit in the traditional sense.

Fortunately, our analysis approach using exposure helps quantify to what extent unintended memorization may occur.

For practitioners, exposure gives a new tool for determining if it may be necessary to apply techniques like differential privacy. Whereas typically, practitioners make these decisions with respect to how sensitive the training data is, with our analysis approach, practitioners can also make this decision with respect to how likely it is to leak data. Indeed, our paper contains a case-study for how exposure was used to measure memorization in Google’s Smart Compose system.

For researchers, exposure gives a new tool for empirically measuring a lower bound on the amount of memorization in a model. Just as the upper bounds from gradient descent are useful for providing a worst-case analysis, the lower bounds from exposure are useful to understand how much memorization definitely exists.


This work was done while the author was a student at UC Berkeley. This article was initially published on the BAIR blog, and appears here with the authors’ permission. We refer the reader to the following paper for details:

Summer travel diary: Reopening cold cases with robotic data discoveries

Traveling to six countries in eighteen days, I journeyed with the goal of delving deeper into the roots of my family before World War II. As a child of refugees, my parents’ narrative is missing huge gaps of information. Still, more than seventy-eight years since the disappearance of my Grandmother and Uncles, we can only presume with a degree of certainty their demise in the mass graves of the forest outside of Riga, Latvia. In our data rich world, archivists are finally piecing together new clues of history using unmanned systems to reopen cold cases.

The Nazis were masters in using technology to mechanize killing and erasing all evidence of their crime. Nowhere is this more apparent than in Treblinka, Poland. The death camp exterminated close to 900,000 Jews over a 15-month period before a revolt led to its dismantlement in 1943. Only a Holocaust memorial stands today on the site of the former gas chamber as a testimony to the memory of the victims. Recently, scientists have begun to unearth new forensic evidence of the Third Reich’s war crimes using LIDAR to expose the full extent of their death factory.

In her work, “Holocaust Archeologies: Approaches and Future Directions,” Dr. Caroline Sturdy Colls undertook an eight-year project to piece together archeological facts from survivor accounts using remote sensors that are more commonly associated with autonomous vehicles and robots than Holocaust studies. As she explains, “I saw working at Treblinka as a cold case where excavation is not permitted, desirable or wanted, [non-invasive] tools offer the possibility to record and examine topographies of atrocity in such a way that the disturbance of the ground is avoided.” Stitching together point cloud outputs from aerial LIDAR sensors, Professor Sturdy Colls stripped away the post-Holocaust vegetation to expose the camp’s original foundations, “revealing the bare earth of the former camp area.” As she writes, “One of the key advantages that LIDAR offers over other remote sensing technologies is its ability to propagate the signal emitted through vegetation such as trees. This means that it is possible to record features that are otherwise invisible or inaccessible using ground-based survey methods.”

Through her research, Sturdy Colls was able to locate several previously unmarked mass graves, transport infrastructure and camp-era buildings, including structures associated with the 1943 prisoner revolt. She credits the technology for her findings, “This is mainly due to developments in remote sensing technologies, geophysics, geographical information systems (GIS) and digital archeology, alongside a greater appreciation of systematic search strategies and landscape profiling,” The researcher stressed the importance of finding closure after seventy-five years, “I work with families in forensics work, and I can’t imagine what it’s like not to know what happened to your family members.” Sturdy Colls’ techniques are now being deployed across Europe at other concentration camp sites and places of mass murder.

Flying north from Poland, I landed in the Netherlands city of Amsterdam to take part in their year-long celebration of Rembrandt (350 years since his passing). At the Rijksmuseum’s Hall of Honors a robot is featured in front of the old master’s monumental work, “Night Watch.” The autonomous macro X-ray fluorescence scanner (Macro-XRF scanner) is busy analyzing the chemical makeup of the paint layers to map and database the age of the pigments. This project, aptly named “Operation Night Watch,” can be experienced live or online showcasing a suite of technologies to determine the best methodologies to return the 1642 painting back to its original glory. Night Watch has a long history of abuse including: two world wars, multiple knifings, one acid attack, botched conservation attempts, and even the trimming of the canvas in 1715 to fit a smaller space. In fact, its modern name is really a moniker of the dirt build up over the years, not the Master’s composition initially entitled: “Militia Company of District II under the Command of Captain Frans Banninck Cocq.”

In explaining the multi-million dollar undertaking the museum’s director, Taco Dibbits, boasted in a recent interview that Operation Night Watch will be the Rijksmuseum’s “biggest conservation and research project ever.” Currently, the Macro-XRF robot takes 24 hours to perform one scan of the entire picture, with a demanding schedule ahead of 56 more scans and 12,500 high-resolution images. The entire project is slated to be completed within a couple of years. Dibbits explains that the restoration will provide insights previously unknown about the painter and his magnum opus: “You will be able to see much more detail, and there will be areas of the painting that will be much easier to read. There are many mysteries of the painting that we might solve. We actually don’t know much about how Rembrandt painted it. With the last conservation, the techniques were limited to basically X-ray photos and now we have so many more tools. We will be able to look into the creative mind of one of the most brilliant artists in the world.”

Whether it is celebrating the narrative of great works of art or preserving the memory of the Holocaust, modern conservatism relies heavily on the accessibility of affordable mechatronic devices. Anna Lopuska, a conservator at the Auschwitz-Birkenau Museum in Poland, describes the Museum’s herculean task, “We are doing something against the initial idea of the Nazis who built this camp. They didn’t want it to last. We’re making it last.” New advances in optics and hardware, enables Lopuska’s team to catalog and maintain the massive camp site with “minimum intervention.” The magnitude of its preservation efforts is listed on its website, which includes: “155 buildings (including original camp blocks, barracks, and outbuildings), some 300 ruins and other vestiges of the camp—including the ruins of the four gas chambers and crematoria at the Auschwitz II-Birkenau site that are of particular historical significance—as well as more than 13 km of fencing, 3,600 concrete fence posts, and many other installations.” This is on top of a collection of artifacts of human tragedy, as each item represents a person, such as “110 thousand shoes, about 3,800 suitcases, 12 thousand pots and pans, 40 kg of eyeglasses, 470 prostheses, 570 items of camp clothing, as well as 4,500 works of art.” Every year more and more survivors pass away making Lopuska’s task, and the unmanned systems she employs, more critical. As the conservationist reminds us, “Within 20 years, there will be only these objects speaking for this place.”

Editor’s Announcements: 1) Vote for our panel, “Love In The Robotic Age,” at SXSW; 2) Signup to attend RobotLab’s next event “Is Today’s Industry 4.0 A Hackers Paradise?” with  Chuck Brooks of General Dynamics on September 25th at 6pm, RSVP Today

#IJCAI2019 main conference in tweets – day 2

Like yesterday, we bring you the best tweets covering major talks and events at IJCAI 2019.

Talks

Paper and Poster Presentations

Demos

50 years old IJCAI panel discussion

Start of industry days


Women’s Lunch

 
Stay tuned as I’ll be covering the conference as an AIhub ambassador.

#IJCAI2019 main conference in tweets


The main IJCAI2019 conference started on August 13th. The organizers gave the opening remarks and statistics, and announced the award winners for this year.

The Opening Ceremony

IJCAI2019 numbers

Special track


Some of the IJCAI2019 Awards

Talks
“Doing for robots what Evolution did for us” by Leslie Kaelbling.

“Human-level intelligence or animal-like abilities” by Adnan Darwiche.

Diversity in AI panel discussion

Demos and booths
Demos and booths of different companies took place next to different poster sessions.

Paper presentation sessions were happening at the same time in other venues.

Robot challenge

 

Stay tuned as I’ll be covering the conference as an AIhub ambassador.

Page 358 of 459
1 356 357 358 359 360 459