Хиймэл оюун ухаан хүний амьдралд нөлөөлөхөд техникийн алдаа болон хүний буруу итгэлцэлтэй холбоотой асуудлууд гарчээ.
2025 оны аравдугаар сарын 20-нд Балтимор хотод 17 настай сурагч Таки Аллен хөлбөмбөгийн дасгалын дараа сургууль дээрээ сууж байхад хиймэл оюун ухаанаар сайжруулсан хяналтын камер түүний халаасанд байсан Doritos чипсийг буу гэж буруу таньжээ. Цагдаагийн автомашинууд ирж, алба хаагчид зэвсэг гарган сурагчийг газарт унагаж, гар хурууг нь холбогдсон байдалтай шалгасан ч зөвхөн чипсний уут олсон байна. Энэ алдаа энгийн оройн уулзалтыг сэтгэл зүйн дарамттай нөхцөл байдал болгож хувиргасан юм.
2025 оны арванхоёрдугаар сарын 24-нд Теннесси мужийн эмэгтэй Ангела Липпс таван сарын турш шоронд хоригдсоны дараа нүүр таних программын алдаанаас болж хойд Дакота мужид гарсан залилангийн хэрэгт буруутгагдсан нь нотлогдсон. Тэрбээр дөрвөн ач зээгээ харж байхад нь буу тулган баривчлагдсан ажээ.
Эдгээр тохиолдлууд нь хиймэл оюун ухааны системүүд магадлал гаргадаг бөгөөд хүмүүс үүнийг баталгаатай үнэн гэж буруу ойлгодогтой холбоотойгоор хүний эрх зөрчигдөж байгааг харуулж байна. Хууль сахиулах байгууллагууд хиймэл оюун ухааныг ашиглахдаа түүний гаргадаг магадлалыг бодитой шийдвэр гэж хүлээж авдаг нь асуудлын гол шалтгаан юм.
АНУ-ын олон хотод хууль сахиулах байгууллагууд хиймэл оюун ухааныг ашиглан гэмт хэргийн түүхэн мэдээлэл дээр үндэслэн эрсдлийн үнэлгээ гаргаж, эрсдэлтэй бүсүүдэд цагдаа нарыг илүү чиглүүлдэг. Гэвч эдгээр системүүдийн гаргадаг статистик мэдээлэл нь тухайн газар орчинд шинэ гэмт хэрэгт холбоотой хүнийг тодорхойлох баталгаа биш юм. Магадлал дээр суурилсан урьдчилсан таамаглал бодит үйл ажиллагааны шийдвэр болж хувирдаг нь хүний эрх зөрчигдөх эрсдлийг нэмэгдүүлдэг.
Хиймэл оюун ухааны загварууд нь бодит баримт биш, харин сургалтанд ашигласан өгөгдлийн загвар дээр үндэслэн хамгийн магадлалтай хариултыг гаргадаг. Жишээ нь, гэрэл асах төхөөрөмжийг хэн зохион бүтээсэн талаар асуухад эдгээр системүүд Томас Эдисон гэж хариулдаг ч энэ нь бүрэн түүхийг илэрхийлэхгүй бөгөөд бусад зохион бүтээгчдийн хувь нэмрийг орхигдуулдаг. Иймээс хиймэл оюун ухааны гаргадаг хариултыг үнэн гэж ойлгох нь эрсдэлтэй.
Хууль сахиулах байгууллагууд хиймэл оюун ухааныг ашиглахдаа магадлалыг баталгаатай гэж үзэхгүйгээр хүний шийдвэртэй хослуулан ашиглах ёстой. Гэхдээ олон байгууллагад ийм бодлого байхгүй бөгөөд системийн гаргадаг мэдэгдлийн итгэлцлийн түвшинг олон нийтэд нууцалдаг. Энэ түвшин нь цагдаагийн үйл ажиллагаа эхлэх эсэхийг тодорхойлдог ч олон нийтэд ил тод биш байна.
Эрүүл мэндийн салбарт ийм төрлийн эрсдэл, алдааг тодорхой хэмжээгээр хүлээн зөвшөөрдөг бөгөөд мэргэжлийн хяналт, ёс зүйн шалгалттай байдаг. Харин хууль сахиулахад хиймэл оюун ухааны гаргадаг алдаатай дохиог хүлээн авах эсэх нь хүний эрх, аюулгүй байдалд шууд нөлөөлдөг тул илүү нарийвчилсан хяналт шаардлагатай.
Хууль эрх зүйн тогтолцоонд нотлох баримтын шаардлага нь хиймэл оюун ухааны магадлалаас ялгаатай бөгөөд шүүхэд итгэлтэй, баталгаатай нотолгоо шаарддаг. Гэтэл хиймэл оюун ухаан ихэвчлэн “итгэлтэй” гэж үзсэн хариултыг гаргадаг ч энэ нь заавал үнэн биш байж болно. Тиймээс технологийг боловсруулж буй мэргэжилтнүүд системийн гаргадаг магадлалыг илэрхийлэх, хэрэглэгчдийг зөв ойлголттой болгох тал дээр анхаарах шаардлагатай байна.
Виржиниа их сургуулийн Digital Technology for Democracy Lab болон бусад судлаачдын үзэж байгаагаар хиймэл оюун ухааныг хууль сахиулахад ашиглахдаа хүний оролцоо, хяналт зайлшгүй шаардлагатай бөгөөд олон нийтийн эрх ашгийг хамгаалах үүднээс ил тод байдал, хариуцлагыг сайжруулах хэрэгтэй гэж үзэж байна.
Дэлгэрэнгүйг эх сурвалжаас харах
↓Эх сурвалжийг нээх ↓
In Baltimore on Oct. 20, 2025, a 17-year-old student named Taki Allen was sitting outside his high school after football practice when an artificial intelligence-enhanced surveillance camera falsely identified the Doritos bag in his pocket as a gun. Within moments police cars arrived, officers drew their weapons and Allen was forced to his knees and handcuffed while they searched him. All they found was a crumpled bag of chips. The AI’s misidentification and the human decisions that followed turned a normal evening into a traumatic confrontation.
On Dec. 24, 2025, Angela Lipps, a Tennessee grandmother, was released after spending five months in jail because facial recognition software had incorrectly connected her to fraud crimes in North Dakota, a state she had never visited. Police had arrested her at gunpoint while she was babysitting her four grandchildren.
These are unfortunate examples of how AI can lead to mistreatment of people because of technical flaws as well as misplaced human faith in the technology’s supposed objectivity. These cases involve different tools, but the underlying issue is the same. AI systems produce probabilities, and people treat them as certainties.
We are researchers who study the intersection of technology, law and public administration. In researching how police departments use AI and how digital technologies operate in a democratic society, we have seen how quickly the shift from probabilistic prediction to operational certainty happens in practice.
AI policing tools are used in dozens of U.S. cities, although no public registry tracks the full footprint. The tools ingest historical crime data and score neighborhoods on predicted risk so officers can be routed toward the resulting hot spots. The mechanism is straightforward, but its consequence is not. Once a system signals a possible threat, the question is no longer how certain the prediction is but what to do about it. A statistical output turns into a deployment decision, and the uncertainty that produced it gets lost on the way.
A matter of probabilities
When generative AI models such as ChatGPT or Claude respond to human requests, they are not searching a database and pulling out facts. They are predicting the most likely answer based on patterns in data they have been trained on. When asked, “Who invented the light bulb?” the models do not go to a source or fact-check a finding. They generate a statistically probable answer which is “Thomas Edison.” The reply might be right, but it might not capture the full story – such as Joseph Swan’s parallel invention at the same time as Edison’s. The danger arises when people believe that the model is retrieving truth rather than generating likelihoods.
This distinction matters. The most probable response is not the same as a factually verified answer, complete with context.
This reality can be highly problematic for policing and law. For example, when law enforcement agencies use AI systems trained on geographical data to estimate where criminal activity is likely to occur, the algorithms analyze historical crime data and geographic patterns. These systems generate statistical risk scores or heat maps for locations based on prior incidents. But such predictions may have little bearing on who was involved in a new crime in the area, even if an algorithm generates information that sounds authoritative.
Some researchers have argued that predictive policing systems do not increase the likelihood that racial minorities will be arrested more often relative to traditional policing practices. The broader concern, however, is not limited to measurable disparities in arrest outcomes alone. It is about how probabilistic predictions can become standardized operational decisions absent further verification.
Artificial intelligence researchers caution against using these models in isolation for crime and legal proceedings or decision-making. Research at the University of Virginia’s Digital Technology for Democracy Lab with police chiefs shows that some law enforcement groups follow strict policies that dictate when technology is used in tandem with, or in place of, human discretion, while others have no such policy.
What most users do not realize is that AI systems rarely produce binary answers: yes or no, a positive identification or a negative one. They generate probabilities. Some systems assign scores that assess the system’s confidence in a prediction. In those cases, engineers set a confidence threshold, a level of certainty that determines when the system should trigger an alert about a possible threat. You can think of this threshold as settings on a control knob. A 95% confidence level, for example, indicates that the model considers its interpretation to be highly likely.
A low threshold catches more potential threats but increases false alarms. A high threshold reduces mistakes but risks missing real dangers. Either way, these algorithmic thresholds are often invisible to the public and are set quietly by vendors or agencies, even though they shape when police action begins.
Where to draw the line
In medicine, these kinds of trade-offs are explicit. Diagnostic tools are calibrated on the relative harm of different errors. In infectious disease settings, for instance, systems that detect infections are often designed to accept more false positives to avoid missing contagious individuals. Then medical professionals look into the human cases. And the algorithm-based decisions are subject to professional standards, ethics reviews and regulatory oversight.
In policing, an AI system must balance false positives, where the system flags a threat that does not exist, and false negatives, where it fails to detect a real danger. The trade-off carries significant consequences. A lower threshold may generate more alerts and allow officers to intervene earlier, but it also increases the risk of mistaken identifications, which happened to Angela Lipps, or escalated encounters like the one Taki Allen experienced. A higher threshold may reduce wrongful interventions but could allow legitimate threats to go undetected.
Some law enforcement agencies argue that acting on imperfect signals is preferable to missing serious risks. But lowering the bar for algorithmic alerts based on probabilistic estimates effectively expands the number of people subjected to police attention. It is important to realize that these thresholds are not neutral features of the technology; they are choices embedded by the creators in the model’s code. Decisions about where to draw the line determine when an algorithmic suspicion becomes a real-world police action, even though the public rarely sees or debates how those thresholds are set.
Limits of optimization
Developers often use several methods to determine where to set a confidence threshold. Techniques such as “receiver operating characteristic curve analysis” examine how changing the threshold for an alert alters the balance between correctly identifying real events and mistakenly flagging harmless ones. Precision–recall analysis examines a similar trade-off, asking how accurate the system’s alerts are relative to the number of incidents it successfully detects.
These approaches could help calibrate systems more responsibly by testing how often an algorithm wrongly flags people or locations. Fine-tuning can improve system performance. But the techniques cannot resolve the underlying question of how much algorithmic uncertainty society is willing to tolerate.
In law, legal standards of proof determine how convincing evidence must be before a judge or jury can rule in favor of a plaintiff or defendant. Courts use formal standards of proof depending on the stakes, such as probable cause, preponderance of the evidence and beyond a reasonable doubt. These standards reflect a societal judgment about how much uncertainty is acceptable before exercising legal authority. A court does not accept a guess or a prediction; it follows a process to weigh evidence. Unlike humans, an AI model does not usually say, “I’m not sure.” A model typically has confidence in its reply, even when the answer is incorrect.
Stakes are rising as AI enters the courtroom, law enforcement, the classroom, the doctor’s office and the public sector. It is important for people to understand that AI does not know things the way many assume it does. It does not distinguish between “maybe” and “definitely.” That is up to us. We believe that technologists should design systems that admit uncertainty and need to educate users about how to interpret AI outputs responsibly.
![]()
Maria Lungu is affiliated with the Digital Technology for Democracy Lab at the University of Virginia, Kennesaw State University, and the Center for DI and Digital Policy (CAIDP).
Steven L. Johnson is affiliated with the Digital Technology for Democracy Lab at the University of Virginia.

