An AI app claims it can detect sexually transmitted infections. Doctors say it's a disaster

Late last month, the San Francisco-based startup HeHealth announced the launch of, a cheerful, emoji-laden website the company describes as “your tech savvy BFF for STI checks.”

The concept is simple. A user concerned about their partner’s sexual health status just snaps a photo (with consent, the service notes) of the partner’s penis (the only part of the human body the software is trained to recognize) and uploads it to Calmara.

In seconds, the site scans the image and returns one of two messages: “Clear! No visible signs of STIs spotted for now” or “Hold!!! We spotted something sus.”

Calmara describes the free service as “the next best thing to a lab test for a quick check,” powered by artificial intelligence with “up to 94.4% accuracy rate” (though finer print on the site clarifies its actual performance is “65% to 96% across various conditions.”)

Since its debut, privacy and public health experts have pointed with alarm to a number of significant oversights in Calmara’s design, such as its flimsy consent verification, its potential to receive child pornography and an over-reliance on images to screen for conditions that are often invisible.

But even as a rudimentary screening tool for visual signs of sexually transmitted infections in one specific human organ, tests of Calmara showed the service to be inaccurate, unreliable and prone to the same kind of stigmatizing information its parent company says it wants to combat.

A Los Angeles Times reporter uploaded to Calmara a broad range of penis images taken from the Centers for Disease Control and Prevention’s Public Health Image Library, the STD Center NY and the Royal Australian College of General Practitioners.

Calmara issued a “Hold!!!” to multiple images of penile lesions and bumps caused by sexually transmitted conditions, including syphilis, chlamydia, herpes and human papillomavirus, the virus that causes genital warts.

Screenshots from the Calmara app with eggplant emoji obscuring photos of genitals.

Screenshots, with genitals obscured by illustrations, show that Calmara gave a “Clear!” to a photo from the CDC of a severe case of syphilis, left, uploaded by The Times; the app said “Hold!!!” on a photo, from the Royal Australian College of General Practitioners, of a penis with no STIs.

(Screenshots via; photo illustration by Los Angeles Times)

But the site failed to recognize some textbook images of sexually transmitted infections, including a chancroid ulcer and a case of syphilis so pronounced the foreskin was no longer able to retract.

Calmara’s AI frequently inaccurately identified naturally occurring, non-pathological penile bumps as signs of infection, flagging multiple images of disease-free organs as “something sus.”

It also struggled to distinguish between inanimate objects and human genitals, issuing a cheery “Clear!” to images of both a novelty penis-shaped vase and a penis-shaped cake.

“There are so many things wrong with this app that I don’t even know where to begin,” said Dr. Ina Park, a UC San Francisco professor who serves as a medical consultant for the CDC’s Division of STD Prevention. “With any tests you’re doing for STIs, there is always the possibility of false negatives and false positives. The issue with this app is that it appears to be rife with both.”

Dr. Jeffrey Klausner, an infectious-disease specialist at USC’s Keck School of Medicine and a scientific adviser to HeHealth, acknowledged that Calmara “can’t be promoted as a screening test.”

“To get screened for STIs, you’ve got to get a blood test. You have to get a urine test,” he said. “Having someone look at a penis, or having a digital assistant look at a penis, is not going to be able to detect HIV, syphilis, chlamydia, gonorrhea. Even most cases of herpes are asymptomatic.”

Calmara, he said, is “a very different thing” from HeHealth’s signature product, a paid service that scans images a user submits of his own penis and flags anything that merits follow-up with a healthcare provider.

Klausner did not respond to requests for additional comment about the app’s accuracy.

Both HeHealth and Calmara use the same underlying AI, though the two sites “may have differences at identifying issues of concern,” co-founder and CEO Dr. Yudara Kularathne said.

“Powered by patented HeHealth wizardry (think an AI so sharp you’d think it aced its SATs), our AI’s been battle-tested by over 40,000 users,” Calmara’s website reads, before noting that its accuracy ranges from 65% to 96%.

“It’s great that they disclose that, but 65% is terrible,” said Dr. Sean Young, a UCI professor of emergency medicine and executive director of the University of California Institute for Prediction Technology. “From a public health perspective, if you’re giving people 65% accuracy, why even tell anyone anything? That’s potentially more harmful than beneficial.”

Kularathne said the accuracy range “highlights the complexity of detecting STIs and other visible conditions on the penis, each with its unique characteristics and challenges.” He added: “It’s important to understand that this is just the starting point for Calmara. As we refine our AI with more insights, we expect these figures to improve.”

On HeHealth’s website, Kularathne says he was inspired to start the company after a friend became suicidal after “an STI scare magnified by online misinformation.”

“Numerous physiological conditions are often mistaken for STIs, and our technology can provide peace of mind in these situations,” Kularathne posted Tuesday on LinkedIn. “Our technology aims to bring clarity to young people, especially Gen Z.”

Calmara’s AI also mistook some physiological conditions for STIs.

The Times uploaded a number of images onto the site that were posted on a medical website as examples of non-communicable, non-pathological anatomical variations in the human penis that are sometimes confused with STIs, including skin tags, visible sebaceous glands and enlarged capillaries.

Calmara identified each one as “something sus.”

Such inaccurate information could have exactly the opposite effect on young users than the “clarity” its founders intend, said Dr. Joni Roberts, an assistant professor at Cal Poly San Luis Obispo who runs the campus’s Sexual and Reproductive Health Lab.

“If I am 18 years old, I take a picture of something that is a normal occurrence as part of the human body, [and] I get this that says that it’s ‘sus’? Now I’m stressing out,” Roberts said.

“We already know that mental health [issues are] extremely high in this population. Social media has run havoc on people’s self image, worth, depression, et cetera,” she said. “Saying something is ‘sus’ without providing any information is problematic.”

Kularathne defended the site’s choice of language. “The phrase ‘something sus’ is deliberately chosen to indicate ambiguity and suggest the need for further investigation,” he wrote in an email. “It’s a prompt for users to seek professional advice, fostering a culture of caution and responsibility.”

Still, “the misidentification of healthy anatomy as ‘something sus’ if that happens, is indeed not the outcome we aim for,” he wrote.

Users whose photos are issued a “Hold” notice are directed to HeHealth where, for a fee, they can submit additional photos of their penis for further scanning.

Those who get a “Clear” are told “No visible signs of STIs spotted for now . . . But this isn’t an all-clear for STIs,” noting, correctly, that many sexually transmitted conditions are asymptomatic and invisible. Users who click through Calmara’s FAQs will also find a disclaimer that a “Clear!” notification “doesn’t mean you can skimp on further checks.”

Young raised concerns that some people might use the app to make immediate decisions about their sexual health.

“There’s more ethical obligations to be able to be transparent and clear about your data and practices, and to not use the typical startup approaches that a lot of other companies will use in non-health spaces,” he said.

In its current form, he said, Calmara “has the potential to further stigmatize not only STIs, but to further stigmatize digital health by giving inaccurate diagnoses and having people make claims that every digital health tool or app is just a big sham.” has raised about $1.1 million since its founding in 2019, co-founder Mei-Ling Lu said. The company is currently seeking another $1.5 million from investors, according to PitchBook.

Medical experts interviewed for this article said that technology can and should be used to reduce barriers to sexual healthcare. Providers including Planned Parenthood and the Mayo Clinic are using AI tools to share vetted information with their patients, said Mara Decker, a UC San Francisco epidemiologist who studies sexual health education and digital technology.

But when it comes to Calmara’s approach, “I basically can see only negatives and no benefits,” Decker said. “They could just as easily replace their app with a sign that says, ‘If you have a rash or noticeable sore, go get tested.’”

Source link

About The Author

Scroll to Top