Connect with us

Boston, MA

‘I couldn’t tell if they were kidding or not.’ 12-year-old wins contest to represent Boston in global LEGO competition. – The Boston Globe

Published

on

‘I couldn’t tell if they were kidding or not.’ 12-year-old wins contest to represent Boston in global LEGO competition. – The Boston Globe


Arlo, Boston’s newly crowned Mini Master Model Builder, spent about two months crafting a replica of the Tea Party scene for LEGO’s first global competition.

His creation, which features underwater and above-water scenes, will be sent to the LEGO House in Billund, Denmark, to vie against 27 other young regional winners from Shanghai to Arizona.

The Global Mini Master Model Builder will be announced in the fall.

“I‘m very excited,” Arlo said in a recent Zoom interview from his home in Dover, N.H., his shaggy blonde hair falling over his forehead. “This means a lot to me because I’ve been building LEGOs for a long time.”

Advertisement
Arlo throws up a peace sign in front of his Boston Tea Party LEGO construction.Lindsay Humphreys

A sixth grader at Dover Middle School, Arlo started playing with LEGOs when he was a toddler and has completed more than 100 LEGO sets.

Still, he never considered entering a LEGO competition until a late February visit to the LEGO Discovery Center Boston in Somerville.

He had decided to make a hand out of LEGOs, not just any hand but “a celestial hand emerging from the depths of our ocean,” when an employee approached him.

Impressed by his work, the employee told Arlo he should enter the regional Boston Mini Master Model Builder contest.

Advertisement

Arlo didn’t know what it was but accepted the challenge.

“So I was like, OK, yeah sure, I’ll enter into the tournament,” he recalled. “And I entered in, and I won.”

Kaleb Thome, the LEGO master builder at the Somerville LEGO center, said he saw Arlo’s hand depiction within a couple of weeks of the deadline for the Boston competition.

“He was one of the last ones to submit,” said Thome, 26. “I immediately knew this might be the one.”

The hand was a “very mature concept,” Thome said, but was “executed really well.” As artists know well, hands are a challenge to portray, Thome said.

Advertisement

“It’s this hand coming up from the ocean,” Thome said. “I was like, dude, that’s a sick concept.”

Arlo and Kaleb Thome pose with his winning LEGO creation. Kaleb Thome

For Arlo, the hand was not a painstaking endeavor; it was simply another opportunity to “create anything,” the reason he was drawn to the building blocks in the first place.

“I think the fact that I could do anything,“ his imagination would allow, Arlo said, explaining his initial interest as a toddler.

When Thome, the judge for the Boston contest, selected Arlo on March 14 over some 50 other applicants across New England, Arlo was shocked.

“I couldn’t tell if they were kidding or not,” he said. “But I was definitely excited, no matter what.”

Advertisement

The win made Arlo Boston’s Mini Master Model Builder and catapulted him into LEGO’s first global competition.

His work had just begun.

From late March to May 31, Arlo spent one to two days a week working in the Somerville LEGO facility on his greatest creation — the Boston Tea Party. His mom or dad would typically drive him down from New Hampshire around noon on Saturday, and Arlo would stay in his workshop until closing.

His Boston Tea Party includes sea creatures, a scene from Spongebob’s Bikini Bottom, and people throwing tea into the harbor. It’s about 10 inches by 20 inches (1 x 2 base plates in LEGO math) and even has a character named Arlo in a red shirt.

Arlo meticulously placed every piece, and Thome said its scale was “the most impressive thing.”

Advertisement

“How much brick and plate he was able to put down,” Thome said. “He really set out some big expectations and goals for himself, and I think he achieved them.”

Besides size, the main rule for global competition is that the building represent the city or region the builder is from.

Arlo, a lover of history LEGO sets, jumped at the opportunity.

“The fact that sometimes they can be like small, little scenes cut into history,” Arlo said of historical LEGO sets. ”So you basically have your own history book without even having to read.”

Arlo said he spent a long time coming up his idea, deciding between Paul Revere’s ride and the Boston Tea Party, he eventually landed on the Harbor scene because of its size. Once he had the idea, he started and didn’t stop.

Advertisement

“If I have an idea, I’m gonna just put it down,” Arlo said. “Once that’s done, I’ll just continue. So I’ll legitimately just build.”

His project will be sent to Denmark after June 30 for the LEGO masters to decide its fate. Until then, it will be on display in Somerville.

The honor of Boston’s Mini Master Builder has required some sacrifices. Arlo didn’t play lacrosse this spring to focus on the LEGO project and has been busy at school, with lot of homework in his favorite subject, science. But his mom, Lindsay Humphreys said it’s been worth it.

“It required the commitment, but knowing that it was such a big deal and probably a once-in-a-lifetime opportunity, for sure, we were supportive of it,” Humphreys said.

Outside of LEGOs, Arlo enjoys video games and playing with his dog Bruno, a beagle and pug mix called a “puggle.”

Advertisement

But LEGOs are his guiding passion.

He wants to be a LEGO master like Thome one day. His dream car, a Volkswagen bus, is from his favorite LEGO set.

Will he get the car one day?

“Maybe if he can get a job and work towards it,” Arlo’s mom said, drawing a grin from Arlo.


Advertisement

Ava Berger can be reached at ava.berger@globe.com. Follow her @Ava_Berger_.





Source link

Boston, MA

Former BYU star Clayton Young crushes lifetime best in Boston — on short notice

Published

on

Former BYU star Clayton Young crushes lifetime best in Boston — on short notice


SALT LAKE CITY — Up until the past month or so, Clayton Young wasn’t sure if he’d make it to the starting line of the 130th Boston Marathon.

By Monday afternoon, he was walking away from the course with a stunning new personal best.

Young finished the 26.2-mile point-to-point course in a personal-record time of 2 hours, 5 minutes and 41 seconds Monday, good for 11th place in an all-time year. Zouhair Talbi ran the fastest time ever by an American, finishing fifth overall in 2:03:45 and Jess McClain broken the American women’s record in 2:20:49.

In all, seven American men and 12 American women finished in the top 20 of the prestigious marathon — including Young, whose streak of six consecutive top-10 finishes dating back to 2023 (including the Paris Olympics) ended, albeit barely.

Advertisement

But donning the No. 24 bib and a brand-new kit for new sponsor Brooks, the former BYU national champion who prepped at American Fork High jumped into the lead pack from the start and never looked back as he broke his previous lifetime best set from the 2023 Chicago marathon and the Olympic trials nearly a year later by close to 3 seconds.

“With only nine weeks of training. … I was really happy to be a 2:05 guy,” Young told FloTrack after the race. “Obviously, falling outside the top 10 is a little disappointing, but I’m really happy with the time.”

The final finish was only the faintest disappointment in the incredibly fast field.

Young’s finish as the third fastest American on Monday marks the fifth-fastest time by an American man all-time in Boston. Charles Hicks finished 50 seconds behind Talbi in 2:04:35, with Young coming in just over a minute later to cheers of friends and family.

His former BYU teammate, Canadian international Rory Linkletter, finished 14th with a personal-best time of 2:06:04. Former BYU runner Michael Ottesen finished 52nd in 2:16:06, and Utah resident Todd Garner finished his 11th running of the Boston Marathon all-time in 3:14:35.

Advertisement

“I think we’re in an era in distance running, on the men and women’s sides, but especially the women’s side, where we’re all making each other so much better every time we line up with one another,” McClain told the Associated Press. “And I think it’s just going to get stronger and stronger.”

Former Utah Valley and BYU runner Kodi Kleven finished 14th in the women’s race with a personal-best time of 2:24:48. The three-time St. George marathon course record holder from Mount Pleasant led for large portions of the race en route to her qualifying time for the 2026 U.S. Olympic marathon trials.

Former BYU standout and Utah State coach Madey Dickson, who also runs trains locally with Run Elite Program, beat her previous personal record in 2:28:12 — good for 18th in the women’s race.

The Key Takeaways for this article were generated with the assistance of large language models and reviewed by our editorial team. The article, itself, is solely human-written.





Source link

Advertisement
Continue Reading

Boston, MA

Tools for Your To Do List with Spot and Gemini Robotics | Boston Dynamics

Published

on

Tools for Your To Do List with Spot and Gemini Robotics | Boston Dynamics


For an industrial robot built for the rigors of factories and power plants, tidying up a living room may seem like a light day at the office for Spot. Yet, a recent video of the robot picking up shoes and soda cans in a residential home represents the promise of AI models in robotics. In this case, Google’s visual-language model (VLM) Gemini Robotics-ER 1.5 was empowering Spot with embodied reasoning.

This particular demo grew out of a 2025 hackathon at Boston Dynamics that built on prior projects using Large Language Models (LLMs) and Visual Foundation Models (VFMs) to enable Spot to contextualize its environment and engage in more complex autonomous actions than a typical Autowalk mission. Rather than write formal software logic or a “state machine” program that defines each step of a given task, we interacted with Gemini Robotics using conversational language. In turn, it communicated with Spot on our behalf.

A Robust SDK and Natural Language Prompts Save Time

Using Spot’s SDK, we developed a layer that facilitated interaction between Gemini Robotics and Spot’s application programming interface (API). The API normally gives developers access to the robot’s capabilities to create custom applications or behaviors. For example, researchers at Meta have used Spot to test how an AI system could locate and retrieve objects it had never seen before.

Advertisement

Our ability to engage Gemini Robotics using natural language prompts was a huge timesaver, compared to traditional programming. We told Gemini Robotics it had access to a mobile robot equipped with cameras and a robotic arm. It also had a finite set of tools it could use to control the robot. A tool is a lightweight script that performs some internal logic and translates inputs from Gemini Robotics to actual API calls. We limited the actions to navigating between locations, capturing images, identifying objects, grasping them, and placing them somewhere else. 

The extent of our SDK means there are great examples one could leverage to add more access to the API with minimal development.

Giving Gemini Robotics a Baseline

To start we needed to explain to Gemini Robotics what we wanted it to do. We did experience a learning curve when writing these baseline prompts. Simple instructions like “put down an object” or “take a picture” weren’t detailed enough to produce expected behavior. We had to add context in our descriptions as we refined each tool. 

A good example is the detailed prompt for the “TakePicture” tool:

This command will cause the robot to take a picture with the specified camera. There is some nuance to choosing the correct camera. Once arriving at a location using GoTo, you should always start by taking a picture with the gripper camera, because it's the most informative.
If the robot has arrived at location and is already holding an object, you can do one of two things:
1. Immediately call PutDown
2. Search the area with either of the front cameras. The front cameras are low to the ground, so if you're trying to put things on an elevated surface, they won't give you useful information.

In this example, we gave Gemini Robotics no detailed description of the robot’s chassis or arm. Instead, we simply explained that Spot’s front cameras would be too low to photograph objects on elevated surfaces. We were able to iterate rapidly, as small changes in wording produced noticeably better results. Once it had this set of basic tools through the API, Gemini Robotics could sequence Spot’s actions and follow the handwritten instructions on a whiteboard on the day of the demonstration.

Advertisement

How Gemini Robotics and Spot Collaborate

Until the robot powers on, Gemini Robotics has no context for what specific tasks we might ask it to perform in a given demo. We only provided simple written instructions, such as, “Make sure all of the shoes at the front door are on the shoe rack.” Gemini Robotics evaluated images from Spot’s cameras and identified objects in the scene that matched the instructions. These objects became the reference points for Spot’s navigational and manipulation systems.

In many respects, Gemini Robotics was identical to an operator manually driving Spot using its tablet controller. For example, to pick up an object with Spot, an operator positions the robot near the object and then uses a grasp wizard to identify the target object. The operator provides high-level direction and Spot figures out the exact details. In this demonstration, Gemini Robotics functioned as both the operator and the tablet sending commands to the robot. This freed us up to act more like a team lead, providing a high-level to-do list and trusting Spot and Gemini Robotics do the rest.

Call and Response

When Gemini Robotics engages a given tool, the tool responds with results and context, such as, “I picked up the object,” or “I can’t pick up something while my hand is full.” Gemini Robotics then makes adjustments on the fly based on this feedback from Spot. For example, to pick up shoes, Gemini Robotics requests an image, identifies the shoes in that image, and calls the “pickup” command. By creating fundamental tools that semantically flow in conversation,  Gemini Robotics can manage the sequence of tasks required to clean up the room. Spot’s existing software stack manages the locomotion, navigation, and manipulation of the robot itself.

It’s important to note Gemini Robotics has strict boundaries in this scenario. It can’t invent new capabilities or control Spot beyond what is available through the API. This keeps Spot’s behavior predictable, while still allowing Gemini Robotics to adapt to different situations.

A Force Multiplier for Developers

For developers already working with Spot, this research has tremendous potential. Through Spot’s SDK, they have access to a robust toolkit of capabilities. Companies use these tools today to build applications for inspection, research, and industrial data analysis, among others.

An AI model like Gemini Robotics offers a way to expand those applications more rapidly. Rather than write extensive task logic on top of Spot’s APIs, developers can experiment with having AI systems interpret natural language instructions and dynamically choose to engage the robot. As a result, models like Gemini Robotics can act as force multipliers, amplifying the reliable toolkit and robust performance that is already delivering value for Boston Dynamics customers.

Advertisement

Our Next-Token Prediction for Spot and Gemini Robotics

Although this is still an experimental step and not a hardened application, it illustrates a compelling direction for robotics and physical AI. Robots like Spot are already extremely capable of navigating complex and changeable environments, collecting data and sensor readings, and manipulating objects. Rather than reinventing the wheel, AI foundation models offer a new way to expand these capabilities in new settings and to new applications.

Physical AI is a rapidly evolving field and our team is leading the way in the lab and in real applications of AI empowered robots. While we are early in our formal partnership with Google Deepmind, we’re excited for what the future holds with Atlas and we’ve already rolled out practical enhancements for Spot and Orbit, with AIVI-Learning powered by Google Gemini Robotics ER 1.6. This next evolution of our AI Visual Inspection tool unlocks a new level of visual intelligence, as users benefit from shared expertise bringing a deeper level of contextual intelligence to Spot and Orbit. Model improvements automatically happen behind the scenes, adding more capabilities to the same software and hardware.

Today, this demo points to a future where users can rely more on natural language to guide Spot’s actions, rather than complex code. The engineer’s role shifts toward setting goals and objectives. The multi-modal robot foundation model interprets the instructions to form complex and adaptive plans and Spot executes the action.

This article was contributed by Issac Ross and Nikhil Devraj, engineers on the Spot team.

Advertisement



Source link

Continue Reading

Boston, MA

A crowd scientist is helping the Boston Marathon manage a growing field of 30,000-plus runners

Published

on

A crowd scientist is helping the Boston Marathon manage a growing field of 30,000-plus runners


BOSTON (AP) — Running the Boston Marathon is tough enough without having to jostle your way from Hopkinton to Copley Square.

So race organizers this year turned to an expert in crowd science to help them manage the field of more than 32,000 as it travels the 26.2 miles (42.195 kilometers) through eight Massachusetts cities and towns — some of it on narrow streets laid out during Colonial times.

“There are certain things that we can’t change — that we don’t want to change — because they make the Boston Marathon,” said Marcel Altenburg, a senior lecturer of crowd science at Manchester Metropolitan University in Britain. “Like, I’m a scientist, but I can’t be too science-y about the race. It should stay what it is because that’s what I love. That’s what the runners love.”

The world’s oldest and most prestigious annual marathon, the Boston race was inspired by the endurance test that made its debut at the inaugural modern Olympics in 1896 — itself a tribute to the route covered by the messenger Pheidippides, who ran to Athens with news of the Greek victory over the Persians in Marathon.

Advertisement

After sharing the news — “Rejoice, we conquer!” — Pheidippides dropped dead.

Organizers of the Boston race would prefer a more pleasant experience for their runners, even as the field has ballooned from 15 in 1897 to as many as 38,000 to meet demand for the 100th edition in 1996. It has settled at around 30,000 since 2015.

As the race grew, it tested the limits of the narrow New England roads and the host cities and towns, which are eager to reopen their streets for regular commutes and commerce as quickly as possible.

“It would be kind of great someday to be able to grow the race a little bit more,” race director Dave McGillivray said. “The problem with this race is that it’s about two things: time and space. We don’t have either. … So, we’re trying to be innovative.”

That’s where Altenburg comes in.

Advertisement

A former German army captain who runs ultra marathons himself, Altenburg has worked with all of the major races, other large sporting events, and airports and exhibitions that tend to attract large crowds on ways to keep things safe and flowing smoothly.

For the Boston Marathon, which draws hundreds of thousands of spectators in addition to the runners, his models allow him to run simulations that help him see how the race might play out under different conditions.

“We have simulated the Boston Marathon more than 100 times to run it once for real. That is the one that counts,” Altenburg said in a telephone interview. “They gave me, pretty much, all creative freedom to simulate more waves, simulate more runners and — within the existing time window — they allowed me to change pretty much anything for the betterment of the running experience.

“And then we checked every aid station, every mile, the finish, every important point, (asking): Is the result better for the runner? Is that something that we should explore further?”

The most noticeable difference on Monday will be that the runners are starting in six waves — groups organized by qualifying time — instead of three. The waves, which were first used in Boston in 2011, help spread things out so that runners don’t have to walk after the start, when Main Street in Hopkinton squeezes to just 39 feet wide.

Advertisement

Other, less obvious changes involve the unloading of the buses at the start, the placement of the water and aid stations, and the finish line chutes, where runners get their medals, perhaps a mylar blanket or a banana, and any medical treatment they might need.

“For an event that’s as old as ours, 130 years, it allowed us to be a startup all over again,” said Lauren Proshan, the chief of race operations and production for the Boston Athletic Association.

“The change isn’t meant to be earth-shattering. It’s to be a smooth experience from start to finish,” she said. “It’s one of those things that you work really, really hard behind the scenes and hope that no one notices — a behind-the-curtain change that makes you feel as if you’re just floating and having a great day.”

Shorter porta potty lines would also be nice.

“What I loved about working with the BAA was how aware they are of what the Boston Marathon is. And they won’t change anything lightly,” Altenburg said. “So it was very detailed work from literally the moment the race last year ended to now. That we check every single option. That we really make sure that if we change something about this historic race, then we know what we’re doing.”

Advertisement

The BAA will look at the feedback over the next three years before deciding about expansion or other changes.

“Fingers crossed, hope for the best, but we’ll get feedback from the participants,” McGillivray said. “And they’ll let us know whether or not it worked or not.”

But keeping the course open longer isn’t an option. And the route isn’t going to change. So there’s only so much that crowd science can help with at one of the toughest tests in sports.

“I can talk. I’m a scientist. I just press a button and it’s going to be,” Altenburg said. “But the runners still have to do it.”

___

Advertisement

AP sports: https://apnews.com/hub/sports



Source link

Continue Reading
Advertisement

Trending