Page 41 of 429
1 39 40 41 42 43 429

Won’t Get Fooled Again

ChatGPT-Generated Exam Answers Dupe Profs

Looks like college take-home tests are destined to suffer the same fate as the Dodo bird.

Instructors at a U.K. university learned as much after a slew of take-home exams featuring answers generated by ChatGPT passed with flying colors — all while evading virtually any suspicions of cheating.

Observes writer Richard Adams: “Researchers at the University of Reading fooled their own professors by secretly submitting AI-generated exam answers that went undetected and got better grades than real students.

“The university’s markers – who were not told about the project – flagged only one of the 33 entries.”

Observes Karen Yeung, a professor at the University of Birmingham: “The publication of this real-world quality assurance test demonstrates very clearly that the generative AI tools — freely and openly available — enable students to cheat take-home examinations without difficulty.”

In other news and analysis on AI writing:

*In-Depth Guide: Lovo AI Text-to-Voice: Writers looking for a reliable text-to-voice solution may want to give Lovo AI a whirl, according to Sharqa Hameed.

Hameed’s guide on the product is extremely valuable in that it offers scores of step-by-step screenshots that truly give you a detailed look at how Lovo AI works.

Hameed’s verdict on the app: “Overall, I’d rate it 4 out of 5.

“It offers various valuable features, including Genny, Auto Subtitle Generator, Text to Speech, Online Video Editor, AI Art Generator, AI Writer and more.

“However, its free version limits you to convert up to 20 minutes of text to audio.”

*Free-for-All: Open Source Promises Wide Array of AI Writing Tools: Facebook founder Mark Zuckerberg predicts that writers and others will continue to have a number of AI choices as the tech grows ever–more sophisticated.

A key player in AI writing/chat tech, Zuckerberg has released his AI code as open-source — available to any and all to use and alter.

Observes Zuckerberg: “I don’t think that AI technology is a thing that should be kind of hoarded and — that one company gets to use it to build whatever central, single product that they’re building.”

*Grand Claims, Meh Results: Google’s AI Falls Short: Apparently, Google’s Gemini — the AI that powers its direct competition to ChatGPT — is not all it’s cracked up to be.

Observes writer Kyle Wiggers: “In press briefings and demos, Google has repeatedly claimed that the models can accomplish previously impossible tasks thanks to their ‘long context.'”

Those tasks include summarizing multiple hundred-page documents or searching across scenes in film footage.

“But new research suggests that the models aren’t, in fact, very good at those things,” Wiggers adds.

*Freelance Writing Dreams Disappearing in a Puff of Code: Add freelancers to the growing list of workers discovering that AI is less a ‘helpful buddy’ and more a ruthless job stealer.

Case in point: Since the advent of ChatGPT, job opportunities in freelance writing have declined 21%, according to a newly updated study.

Observes writer Laura Bratton: “Research shows that easily-automated writing and (computer) coding jobs are being replaced by AI.”

*Privacy Ninja: New AI Email Promises to Guard Your Secrets: Proton, an email provider long-prized for its heavy emphasis on privacy, has added AI to its mix.

Specifically, its newly released AI writing assistant ‘Proton Scribe’ is designed to help users auto-write and proofread their emails.

Observes writer Paul Sawers: “Proton Scribe can be deployed entirely at the local device level — meaning user data doesn’t leave the device.

“Moreover, Proton promises that its AI assistant won’t learn from user data — a particularly important feature for enterprise use cases, where privacy is paramount.”

*Forget Solitaire: Claude Turns AI Writing into a Collaborative Party Game: ChatGPT competitor Claude has a new feature that enables users to publish, share and remix the AI writing and other content that they generate with one another.

Observes writer Eric Hal Schwartz: “Essentially, you can open published ‘Artifacts’ created by others and modify or build upon them through conversations with Claude.

“Anthropic is pitching it as a way to foster a collaborative environment.”

*Robo Lawyer: For Many Attorneys, AI Still a Boogeyman: Despite its considerable benefits to the legal community, AI is viewed warily by many lawyers and pros.

Specifically, 77% of professionals recently surveyed by Thomson Reuters saw AI as a threat to lawyers.

Observes Artificial Lawyer: “While some very innovative lawyers are comfortable with AI and have few worries about the legal world’s imminent demise, there are plenty of lawyers out there who still feel very uncertain about what this all means for them, the profession, and their firms.”

*ChatGPT Mind-Meld: New Hope For the Paralyzed: A man slowly succumbing to paralysis has been given new hope with ChatGPT, which is enabling him to text using his brain waves.

Using a brain implant, the man is able to translate his thoughts into text commands — generated by ChatGPT — which he uses to operate computerized communication devices.

Observes the patient: “You get choices of how you might respond in several different ways.

“So rather than me typing single words, I’m hitting one or two buttons or clicks, if you will, and I’ve got the majority of a sentence done.”

*AI Big Picture: AI Gold Rush Still Runs Hot: Nearly 20 months after ChatGPT introduced a stunned world to the potential of AI, businesses across the world are still clamoring to bring the newly commercialized tech onboard.

Observes writer Ben Dickson: “Most organizations are spending hefty amounts to either explore generative AI use cases or have already implemented them in production,” according to a new survey of 200 IT leaders.

“Nearly three-fourths (73%) of respondents plan to spend more than $500,000 on generative AI in the next 12 months, with almost half (46%) allocating more than $1 million,” Dickson adds.

Share a Link:  Please consider sharing a link to https://RobotWritersAI.com from your blog, social media post, publication or emails. More links leading to RobotWritersAI.com helps everyone interested in AI-generated writing.

Joe Dysart is editor of RobotWritersAI.com and a tech journalist with 20+ years experience. His work has appeared in 150+ publications, including The New York Times and the Financial Times of London.

Never Miss An Issue
Join our newsletter to be instantly updated when the latest issue of Robot Writers AI publishes
We respect your privacy. Unsubscribe at any time -- we abhor spam as much as you do.

The post Won’t Get Fooled Again appeared first on Robot Writers AI.

#RoboCup2024 – daily digest: 21 July

A break in play during a Small Size League match.

Today, 21 July, saw the competitions draw to a close in a thrilling finale. In the third and final of our round-up articles, we provide a flavour of the action from this last day. If you missed them, you can find our first two digests here: 19 July | 20 July.

My first port of call this morning was the Standard Platform League, where Dr Timothy Wiley and Tom Ellis from Team RedbackBots, RMIT University, Melbourne, Australia, demonstrated an exciting advancement that is unique to their team. They have developed an augmented reality (AR) system with the aim of enhancing the understanding and explainability of the on-field action.

The RedbackBots travelling team for 2024 (L-to-R: Murray Owens, Sam Griffiths, Tom Ellis, Dr Timothy Wiley, Mark Field, Jasper Avice Demay). Photo credit: Dr Timothy Wiley.

Timothy, the academic leader of the team explained: “What our students proposed at the end of last year’s competition, to make a contribution to the league, was to develop an augmented reality (AR) visualization of what the league calls the team communication monitor. This is a piece of software that gets displayed on the TV screens to the audience and the referee, and it shows you where the robots think they are, information about the game, and where the ball is. We set out to make an AR system of this because we think it’s so much better to view it overlaid on the field. What the AR lets us do is project all of this information live on the field as the robots are moving.”

The team has been demonstrating the system to the league at the event, with very positive feedback. In fact, one of the teams found an error in their software during a game whilst trying out the AR system. Tom said that they’ve received a lot of ideas and suggestions from the other teams for further developments. This is one of the first (if not, the first) AR system to be trialled across the competition, and first time it has been used in the Standard Platform League. I was lucky enough to get a demo from Tom and it definitely added a new level to the viewing experience. It will be very interesting to see how the system evolves.

Mark Field setting up the MetaQuest3 to use the augmented reality system. Photo credit: Dr Timothy Wiley.

From the main soccer area I headed to the RoboCupJunior zone, where Rui Baptista, an Executive Committee member, gave me a tour of the arenas and introduced me to some of the teams that have been using machine learning models to assist their robots. RoboCupJunior is a competition for school children, and is split into three leagues: Soccer, Rescue and OnStage.

I first caught up with four teams from the Rescue league. Robots identify “victims” within re-created disaster scenarios, varying in complexity from line-following on a flat surface to negotiating paths through obstacles on uneven terrain. There are three different strands to the league: 1) Rescue Line, where robots follow a black line which leads them to a victim, 2) Rescue Maze, where robots need to investigate a maze and identify victims, 3) Rescue Simulation, which is a simulated version of the maze competition.

Team Skollska Knijgia, taking part in the Rescue Line, used a YOLO v8 neural network to detect victims in the evacuation zone. They trained the network themselves with about 5000 images. Also competing in the Rescue Line event were Team Overengeniering2. They also used YOLO v8 neural networks, in this case for two elements of their system. They used the first model to detect victims in the evacuation zone and to detect the walls. Their second model is utilized during line following, and allows the robot to detect when the black line (used for the majority of the task) changes to a silver line, which indicates the entrance of the evacuation zone.

Left: Team Skollska Knijgia. Right: Team Overengeniering2.

Team Tanorobo! were taking part in the maze competition. They also used a machine learning model for victim detection, training on 3000 photos for each type of victim (these are denoted by different letters in the maze). They also took photos of walls and obstacles, to avoid mis-classification. Team New Aje were taking part in the simulation contest. They used a graphical user interface to train their machine learning model, and to debug their navigation algorithms. They have three different algorithms for navigation, with varying computational cost, which they can switch between depending on the place (and complexity) in the maze in which they are located.

Left: Team Tanorobo! Right: Team New Aje.

I met two of the teams who had recently presented in the OnStage event. Team Medic’s performance was based on a medical scenario, with the team including two machine learning elements. The first being voice recognition, for communication with the “patient” robots, and the second being image recognition to classify x-rays. Team Jam Session’s robot reads in American sign language symbols and uses them to play a piano. They used the MediaPipe detection algorithm to find different points on the hand, and random forest classifiers to determine which symbol was being displayed.

Left: Team Medic Bot Right: Team Jam Session.

Next stop was the humanoid league where the final match was in progress. The arena was packed to the rafters with crowds eager to see the action.
Standing room only to see the Adult Size Humanoids.

The finals continued with the Middle Size League, with the home team Tech United Eindhoven beating BigHeroX by a convincing 6-1 scoreline. You can watch the livestream of the final day’s action here.

The grand finale featured the winners of the Middle Size League (Tech United Eindhoven) against five RoboCup trustees. The humans ran out 5-2 winners, their superior passing and movement too much for Tech United.

Landmark Study Reveals Wearable Robotics Significantly Boost Safety and Efficiency in Industrial Environments

The four-year study, the first of its kind, tracked ergonomics, safety, and injury metrics across more than 65 million lifts at over 40 sites in five industries: Construction, Food & Beverage, Logistics, Manufacturing, and Retail.

#RoboCup2024 – daily digest: 20 July

The Standard Platform Soccer League in action.

This is the second of our daily digests from RoboCup2024 in Eindhoven, The Netherlands. If you missed the first digest, which gives some background to RoboCup, you can find it here.

Competitions continued across all the leagues today, with participants vying for a place in Sunday’s finals.

The RoboCup@Work league focusses on robots in work-related scenarios, utilizing ideas and concepts from other RoboCup competitions to tackle open research challenges in industrial and service robotics.

I arrived at the arena in time to catch the advanced navigation test. Robots have to autonomously navigate, picking up and placing objects at different work stations. In this advanced test, caution tape is added to the arena floor, which the robots should avoid travelling over. There is also a complex placing element where teams have to put an object that they’ve collected into a slot – get the orientation or placement of the object slightly wrong and the it won’t fall into the slot.

The RoboCup@Work arena just before competition start.

Eight teams are taking part in the league this year. Executive Committee member Asad Norouzi said that there are plans to introduce a sub-league which would provide an entry point for new teams or juniors to get into the league proper.

I caught up with Harrison Burns, Mitchell Torok and Jasper Arnold from Team MiRobot. They are based at the University of New South Wales and are attending RoboCup for the first time.

Team MiRobot from UNSW.

The team actually only started six months ago, so final preparations have been a bit stressful. However, the experience has been great fun, and the competition has gone well so far. Like most teams, they’ve had to make many refinements as the competition has progressed, leading to some late nights.

One notable feature of the team’s robot is the bespoke, in-house-designed grasping mechanism on the end of the arm. The team note that “it has good flexible jaws, so when it grabs round objects it actually pulls the object directly into it. Because it uses a linear motion, compared to a lot of other rotating jaws, it has a lot better reliability for picking up objects”.

Here is some footage from the task, featuring Team bi-t-bots and Team Singapore.

In the Middle Size Soccer league (MSL), teams of five fully autonomous robots play with a regular size FIFA ball. Teams are free to design their own hardware but all sensors have to be on-board and there is a maximum size and weight limit of 40kg for the robots. The research focus is on mechatronics design, control and multi-agent cooperation at plan and perception levels. Nine teams are competing this year.

I spoke to António Ribeiro, who is a member of the technical committee and part of Team LAR@MSL from the University of Minho, Portugal. The team started in 1998, but António and most of his colleagues on the current team have only been involved in the MSL since September 2022. The robots have evolved as the competition has progressed, and further improvements are in progress. Refinements so far have included communication, the detection system, and the control system. They are pleased with the improvements from the previous RoboCup. “Last year we had a lot of hardware issues, but this year the hardware seems pretty stable. We also changed our coding architecture and it is now much easier and faster for us to develop code because we can all work on the code at the same time on different modules”.

António cited versatility and cost-effective solutions as strengths of the team. “Our robot is actually very cheap compared to other teams. We use a lot of old chassis, and our solutions always go to the lowest cost possible. Some teams have multiple thousand dollar robots, but, for example, our vision system is around $70-80. It works pretty well – we need to improve the way we handle it, but it seems stable”.

Team LAR@MSL

The RoboCup@Home league aims to develop service and assistive robot technology with high relevance for future personal domestic applications. A set of benchmark tests is used to evaluate the robots’ abilities and performance in a realistic non-standardized home environment setting. These tests include helping to prepare breakfast, clearing the table, and storing groceries.

I arrived in time to watch the “stickler for the rules” challenge, where robots have to navigate different rooms and make sure that the people inside (“guests” at a party) are sticking to four rules: 1) there is one forbidden room – if a guest is in there the robot must alert them and ask them to follow it into another room), 2) everyone must have a drink in their hand – if not, the robot directs them to a shelf with drinks, 3) no shoes to be worn in the house, 4) there should be no rubbish left on the floor.

After watching an attempt from the LAR@Home robot, Tiago from the team told me a bit about the robot. “The goal is to develop a robot capable of multi general-purpose tasks in home and healthcare environments.” With the exception of the robotic arm, all of the hardware was built by the team. The robot has two RGBD cameras, two LIDARs, a tray (where the robot can store items that it needs to carry), and two emergency stop buttons that deactivate all moving parts. Four omnidirectional wheels allow the robot to move in any direction at any time. The wheels have independent suspension systems which guarantees that they can all be on the ground at all times, even if there are bumps and cables on the venue floor. There is a tablet that acts as a visual interface, and a microphone and speakers to enable communication between humans and the robot, which is all done via speaking and listening.

Tiago told me that the team have talked to a lot healthcare practitioners to find out the main problems faced by elderly people, and this inspired one of their robot features. “They said that the two main injury sources are from when people are trying to sit down or stand up, and when they are trying to pick something up from the floor. We developed a torso that can pick objects from the floor one metre away from the robot”.

The LAR@Home team.


You can keep up with the latest news direct from RoboCup here.

Click here to see all of our content pertaining to RoboCup.

Drones could revolutionize the construction industry, supporting a new UK housing boom

The UK's new Labor government has pledged to build 1.5 million homes during its first term in office. To achieve this, it promises planning reform and the reintroduction of local housing targets. Yet little attention has been given to the role of new technologies.

Are We Ready for Multi-Image Reasoning? Launching VHs: The Visual Haystacks Benchmark!

Humans excel at processing vast arrays of visual information, a skill that is crucial for achieving artificial general intelligence (AGI). Over the decades, AI researchers have developed Visual Question Answering (VQA) systems to interpret scenes within single images and answer related questions. While recent advancements in foundation models have significantly closed the gap between human and machine visual processing, conventional VQA has been restricted to reason about only single images at a time rather than whole collections of visual data.

This limitation poses challenges in more complex scenarios. Take, for example, the challenges of discerning patterns in collections of medical images, monitoring deforestation through satellite imagery, mapping urban changes using autonomous navigation data, analyzing thematic elements across large art collections, or understanding consumer behavior from retail surveillance footage. Each of these scenarios entails not only visual processing across hundreds or thousands of images but also necessitates cross-image processing of these findings. To address this gap, this project focuses on the “Multi-Image Question Answering” (MIQA) task, which exceeds the reach of traditional VQA systems.


Visual Haystacks: the first "visual-centric" Needle-In-A-Haystack (NIAH) benchmark designed to rigorously evaluate Large Multimodal Models (LMMs) in processing long-context visual information.

Read More

#RoboCup2024 – daily digest: 19 July

The main soccer arena.

RoboCup is an international scientific initiative with the goal to advance the state of the art of intelligent robots. As part of this initiative, a series of competitions and events are held throughout the year. The main showcase event is an international affair with teams travelling from far and wide to put their machines through their paces.

This year, RoboCup is being held in three arenas in the Genneper Parken, Eindhoven, The Netherlands. The organisers are expecting over 2,000 participants, from 45 different countries, with around 300 teams signed up to take part in the various competitions.

Although RoboCup started out as a football (or soccer) playing competition, other leagues have since been introduced, focussing on robots in industrial, rescue, and home settings. There is even a dedicated league for young roboticists – RoboCupJunior – where participants can take part in either football, rescue, or artistic events.

I am lucky enough to be able to attend this year, and, for the next three days, I’ll be bringing you a daily digest of some of the exciting happenings from Eindhoven.

Today, 19 July, sees the competition in full swing. The main soccer arena, boasting multiple pitches, hosts a number of the different leagues which form RoboCupSoccer.

Some of the pitches in the main soccer arena.

My first port of call was the Standard Platform League, where the round 5 champions cup match between SPQR Team vs rUNSWift was taking place. SPQR ran out winners and advance to round 6. In this league, all teams compete with identical robots (currently the humanoid NAO by Aldebaran). The robots operate fully autonomously, meaning that there is no external control from neither humans nor computers.

Standard platform league. Round 5 champions cup match between SPQR Team vs rUNSWift.

The Humanoid AdultSize league is arguably the most challenging of the leagues, with many constraints placed on the robots to make them as human-like as possible. For example, they must have roughly human-like body proportions, they need to walk on two legs, and they are only allowed to use human-like sensors (up to two cameras to sense the environment). In this AdultSize competition, two robots from each team compete, and the team members walk behind the robots to catch them in case of a fall. Such a mishap could prove costly in terms of potential hardware damage.

Action from the Humanoid AdultSize League.

The RoboCup Rescue Robot League sees teams developing robotic systems with the goal of enabling emergency responders to perform extremely hazardous tasks from safer stand-off distances. During the competition, teams compete in a round-robin, putting their robots through their paces on a number of different challenges. The leading teams following this initial phase progress to the finals on Sunday. The tasks include navigating in complex environments, opening doors, and sensing. Teams may run the machines completely autonomously, or with some assistive control. More points are awarded for completely autonomous operation.

RoboCup Rescue arena from above.

You can keep up with more RoboCup2024 news here.

Researchers use light to control ferrofluid droplet movements in water

A team of engineers at the Max Planck Institute for Intelligent Systems, the Chinese University of Hong Kong and the Gwangju Institute of Science and Technology has found that ferrofluidic drops in a tank of water can be forced to rise in desired ways using light. The study is published in the journal Science Advances.

IBM Uses AI for the Next Generation of Support

This month, IBM had an interesting briefing on how AI could be used to improve the customer support experience, drive deeper engagement with customers, and effectively improve customer loyalty while dramatically reducing support costs. This sounds unusually good given how […]

The post IBM Uses AI for the Next Generation of Support appeared first on TechSpective.

How a S.E.A of Data Unlocks AI’s Potential in Recruitment

The emergence of artificial intelligence is elevating operational efficiency for organizations worldwide. Conference keynotes and media headlines are dominated by the latest AI advancements, generating intrigue and excitement. While the full potential of AI remains unknown, organizations across all industries […]

The post How a S.E.A of Data Unlocks AI’s Potential in Recruitment appeared first on TechSpective.

New framework allows robots to learn via online human demonstration videos

To be successfully deployed in real-world settings, robots should be capable of reliably completing various everyday tasks, ranging from household chores to industrial processes. Some of the tasks they could complete entail manipulating fabrics, for instance when folding clothes to put them in a wardrobe or helping older adults with mobility impairments to knot their ties before a social event.
Page 41 of 429
1 39 40 41 42 43 429