Signal Versus Noise: Evaluating iNaturalist Photos as a Source of Quantitative Phenotypic Data in Plethodon Salamanders using Autoresearch and Agentic AI

This study demonstrates that while iNaturalist photographs can effectively recover discrete color morph frequencies in *Plethodon* salamanders, they are unsuitable for continuous quantitative phenotyping like dorsal brightness due to overwhelming observer-induced noise that obscures geographic signals.

O'Connell, K. A.

Published 2026-03-27
📖 5 min read🧠 Deep dive
⚕️

This is an AI-generated explanation of a preprint that has not been peer-reviewed. It is not medical advice. Do not make health decisions based on this content. Read full disclaimer

The Big Picture: Can We Learn from Random Photos?

Imagine you have a massive library of millions of photos taken by regular people (citizen scientists) of salamanders in the wild. These photos are tagged with exactly where they were taken. Scientists want to use these photos to answer a big question: Do salamanders get darker or lighter depending on where they live?

This is a classic biology question. Some theories say salamanders in hot, humid places should be darker (to handle the sun), while others say those in cold places should be darker (to soak up heat).

The author of this paper, Kyle O'Connell, decided to test if these "messy" crowd-sourced photos are good enough to answer this question. He treated the photos like a signal trying to get through a wall of static (noise).

The Problem: The "Camera Settings" Wall

The main issue with these photos is that they aren't taken in a lab. One person might take a photo with a flash on a sunny day; another might take one with a phone in the shade at night.

The Analogy: Imagine trying to measure the exact temperature of a room by asking 10,000 people to guess the temperature using their own broken thermometers. Some thermometers are stuck on "Hot," some on "Cold," and some are just broken. Even if the room temperature actually changes slightly from one end of the house to the other, your data will look like random chaos because the people and their tools are the biggest source of error, not the room itself.

The Experiment: The "Robot Researcher"

To fix this, the author didn't just manually tweak the computer code. He used a new, fancy method called "Autoresearch" (powered by AI).

The Analogy: Think of this like a robot chef trying to make the perfect soup. Instead of the chef guessing, the robot tries 50 different recipes in rapid succession. It changes one thing at a time (a pinch more salt, a different pot, a hotter stove), tastes the soup, and keeps the changes that make it better.

  • The "soup" was the computer code that measures salamander color.
  • The "taste test" was checking if the code could find a pattern in the data.

The robot tried everything: cropping the photos differently, changing how it calculated color, and filtering out bad images.

The Results: Two Different Stories

The study looked at two types of salamander traits:

1. The Continuous Trait: "How Dark is the Salamander?" (The Failure)

The author tried to measure the exact shade of gray or brown on the salamander's back.

  • The Result: Total failure. The computer found almost no connection between the salamander's color and its location.
  • Why? The "noise" was too loud. The study found that who took the photo explained 23% of the color differences. If Person A took a photo, the salamander looked bright; if Person B took a photo of the same salamander, it looked dark.
  • The Takeaway: You cannot use these random photos to measure exact shades of color. The "broken thermometers" (camera settings) are too broken to detect the tiny temperature changes (biological color shifts).

2. The Discrete Trait: "Is it Red or Gray?" (The Success)

The author then tried a simpler question: Is the salamander the "Red-Back" variety or the "Lead-Back" (gray) variety? This is a yes/no question, not a "how much" question.

  • The Result: Success! The computer could find a geographic pattern. It found that red-backed salamanders were slightly more common in certain areas.
  • Why? The difference between "Red" and "Gray" is so huge that even a bad camera can tell them apart. It's like trying to tell the difference between a red ball and a gray rock in the dark; even with bad eyesight, you can still see the difference.
  • The Catch: The study also found that people tend to take photos of the "weird" looking salamanders (the rare ones) more often than the common ones. So, while the computer could find the pattern, the pattern might be slightly distorted because people prefer taking pictures of the "cool" looking ones.

The "Crop Quality" Reality Check

The author also did a manual check. He looked at 200 photos the computer thought were "good."

  • The Shock: Only 38% were actually good!
  • Many photos were blurry, taken from the wrong angle, or showed the salamander being held in a human hand (which ruins the color reading).
  • The Irony: The computer's automatic "quality check" passed almost all of these bad photos. It was like a bouncer at a club who let in everyone, even people wearing pajamas, because they looked "okay" from a distance.

The Final Verdict

What can we learn from these photos?

  • Can we measure exact colors? No. The photos are too messy. The "signal" (biology) is drowned out by the "noise" (bad cameras and lighting).
  • Can we count different types? Yes, but with caution. If the difference is big (Red vs. Gray), we can see it. But we have to be careful because people might be taking more photos of the rare types, skewing the numbers.

The Big Lesson for Science:
Before scientists spend years analyzing millions of crowd-sourced photos, they should run a "robot chef" test first. This study shows that for some questions (exact measurements), these photos are useless. For others (big categories), they are useful, but you have to be very careful about how you ask the question.

In short: You can use a crowd-sourced photo album to tell if a salamander is wearing a red shirt or a gray shirt, but you can't use it to measure the exact shade of red or the precise brightness of the gray. The "noise" of the photographers is just too loud.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →