What are the markers of trust for generative AI?

On 30th Novem­ber 2022, with the launch of ChatGPT to the gene­ral public1, gene­ra­tive AI left the labo­ra­to­ry and ente­red mee­ting rooms, finan­cial ser­vices, hos­pi­tals, schools, and more. The main advan­tage of this tech­no­lo­gy is well known – with just a few clicks, it can trans­form a mass of data into fluid, intel­li­gible text. Today, with this tool, a finan­cial direc­tor can obtain an auto­ma­tic com­men­ta­ry on his mar­gins in a mat­ter of seconds, a doc­tor can obtain a report based on exa­mi­na­tions, and a student can gene­rate an essay from a simple statement.

This conve­nience and ease of use are a game-chan­ger. Where busi­ness intel­li­gence main­ly pro­du­ced figures and graphs, gene­ra­tive models add a layer of inter­pre­ta­tion. They prio­ri­tise signals, offer expla­na­tions and some­times sug­gest fore­casts. Howe­ver, a clear nar­ra­tive gives the impres­sion of obvious­ness : the conclu­sion seems robust because it is well for­mu­la­ted, even though it is based on just one model among many2.

The risk lies not in the use of AI, but in the exces­sive cre­di­bi­li­ty given to texts for which we often do not know the condi­tions of pro­duc­tion. In other words, can we decide on an invest­ment of seve­ral mil­lion pounds or make a medi­cal diag­no­sis based on the recom­men­da­tions and inter­pre­ta­tions of gene­ra­tive AI ?

Re-examining the trust given

The trust given to a nume­ri­cal res­ponse is usual­ly based on two condi­tions : the qua­li­ty of the source data and the trans­pa­ren­cy of the cal­cu­la­tion method. Howe­ver, in the case of a lite­ral res­ponse such as that pro­du­ced with gene­ra­tive AI, a third layer is added : the inter­pre­ta­tion of the model3.

Indeed, the model decides what to high­light, dis­cards cer­tain ele­ments and impli­cit­ly com­bines variables. The final pro­duct is an auto­ma­ted nar­ra­tive that bears the mark of invi­sible sta­tis­ti­cal and lin­guis­tic choices. These choices may be rela­ted to the fre­quen­cy of the data used to build the model, pro­blem-sol­ving methods or any other cause. To ensure confi­dence in the ans­wer given, these steps should be audi­table, i.e. indi­ca­ted by the user, who can then veri­fy them.

It is now impor­tant to ima­gine a form of algo­rith­mic audit, no lon­ger just veri­fying data but control­ling the entire chain 

This solu­tion, which allows for veri­fi­ca­tion, alrea­dy exists in simi­lar situa­tions. First of all, sho­wing the thought pro­cess is a com­mon approach in tea­ching mathe­ma­tics, as it allows the tea­cher to ensure that the student has unders­tood the steps invol­ved in the rea­so­ning. Simi­lar­ly, in finan­cial ana­ly­sis, audits are used to veri­fy com­pliance with accoun­ting rules. Finan­cial audits gua­ran­tee that the publi­shed figures cor­res­pond to a mea­su­rable reality.

Thus, it is now neces­sa­ry to ima­gine a form of “algo­rith­mic audit”: no lon­ger just veri­fying data but control­ling the entire chain that leads from the raw flow to the final nar­ra­tive. Take the example of a hos­pi­tal where gene­ra­tive AI sum­ma­rises patient records. If it sys­te­ma­ti­cal­ly omits cer­tain cli­ni­cal para­me­ters dee­med rare, it pro­duces attrac­tive but incom­plete reports. The audit must the­re­fore test the robust­ness of the model, assess its abi­li­ty to repro­duce aty­pi­cal cases and veri­fy the tra­cea­bi­li­ty of sources. Simi­lar­ly, an auto­ma­tic ener­gy report that ignores abnor­mal consump­tion peaks can give a false impres­sion of sta­bi­li­ty. Here again, the audit must ensure that ano­ma­lies are taken into account.

Technical protocols to be optimised and deployed more widely

Trust-based engi­nee­ring can­not rely sole­ly on decla­ra­tions of prin­ciple. It must be trans­la­ted into spe­ci­fic pro­to­cols. A num­ber of approaches are alrea­dy emerging :

  • Tra­cea­bi­li­ty of algo­rith­mic choices : each indi­ca­tor must be lin­ked to the source data and the pro­ces­sing applied. This involves docu­men­ting trans­for­ma­tions, as we cur­rent­ly docu­ment a sup­ply chain. Cir­cuit tra­cing methods can pro­vide tra­cea­bi­li­ty that is unders­tan­dable to humans4. Tra­cea­bi­li­ty then becomes an edu­ca­tio­nal tool as well as a control mechanism.
  • Model stress tests : expo­sing AI to unu­sual sce­na­rios to mea­sure its abi­li­ty to reflect uncer­tain­ty rather than smooth it out. For example, it is very use­ful to use a sample that does not fol­low the clas­sic dis­tri­bu­tion to check whe­ther the deep AI model has been inte­gra­ted, inde­pen­dent­ly of the test set pro­vi­ded5. This could involve pro­vi­ding a set of lung X‑rays from smo­kers only. This makes it pos­sible to veri­fy that the AI does not gene­rate an excess of ‘false nega­tives’ to return to a sta­tis­ti­cal average.
  • Mini­mum explai­na­bi­li­ty gua­ran­teed : Without revea­ling the algo­rith­mic secrets of com­pa­nies pro­vi­ding AI solu­tions, it is envi­sa­ged that they will be asked to pro­vide at least a sum­ma­ry of the main variables used in their models to reach a conclu­sion. This explai­na­bi­li­ty could be sub­ject to either ISO-type cer­ti­fi­ca­tion for AI qua­li­ty or vali­da­tion by a regu­la­to­ry body (pre­fe­ra­bly an exis­ting one so as not to mul­ti­ply the num­ber of authorities).

These methods will not remove the confi­den­tia­li­ty asso­cia­ted with the spe­ci­fic and dif­fe­ren­tia­ting set­tings of the manu­fac­tu­rers who deve­lop large lan­guage models, but they will reduce the risk of blind­ness and unjus­ti­fied confi­dence. The issue is not to make AI com­ple­te­ly trans­pa­rent, but to create suf­fi­cient safe­guards to main­tain trust.

An organisational culture in need of transformation

Beyond the tech­ni­cal dimen­sion, it is neces­sa­ry to pro­mote a major cultu­ral shift. For decades, orga­ni­sa­tions have been accus­to­med to vie­wing figures as cer­tain­ties. Dash­boards are often per­cei­ved as indis­pu­table truths. With gene­ra­tive AI and its exten­sion to lite­ral and sub­jec­tive pro­duc­tions, this stance is beco­ming untenable.

Deci­sion-makers, as well as all digi­tal sta­ke­hol­ders, must learn to read an auto­ma­tic report as a sta­tis­ti­cal res­ponse based on known or unk­nown assump­tions, and above all not as a defi­ni­tive conclu­sion. This means trai­ning users of AI solu­tions to for­mu­late deman­ding requests (asking for the AI’s ‘rea­so­ning’ pro­cess) and to read res­ponses cri­ti­cal­ly : iden­ti­fying mar­gins of error, ques­tio­ning omis­sions, and asking for alter­na­tive sce­na­rios. In other words, rein­tro­du­cing uncer­tain­ty into the very heart of the deci­sion-making process.

The Euro­pean Union has begun to lay the ground­work with the AI Act, which clas­si­fies the use of AI in finance and public gover­nance as ‘high risk’. This regu­la­tion imposes an obli­ga­tion of trans­pa­ren­cy and audi­ta­bi­li­ty. But the law will not be enough if orga­ni­sa­tions do not culti­vate active vigi­lance. Gene­ra­tive AI must be control­led not only by stan­dards, but also by a dai­ly prac­tice of cri­ti­cal reading.

Moving towards a measure of vigilance

Gene­ra­tive AI is nei­ther a mirage nor a pana­cea. It speeds up access to infor­ma­tion and pro­vides cla­ri­ty on volumes of data that are unma­na­geable for humans, but it also trans­forms our rela­tion­ship with deci­sion-making. Where we used to see num­bers, we now read stories.

The chal­lenge is the­re­fore not to turn back the clock, but to invent a new engi­nee­ring of trust. Tra­cea­bi­li­ty of cal­cu­la­tions, stress tests, mini­mal explai­na­bi­li­ty : these are all tech­ni­cal buil­ding blocks that need to be put in place, bea­ring in mind that an AI model is like­ly to be the tar­get of mul­tiple cybe­rat­tacks 6,7.

But the key lies in adop­ting a new orga­ni­sa­tio­nal culture : accep­ting that uncer­tain­ty is a given and not a fai­lure of the sys­tem. Only then can gene­ra­tive AI become a reliable tool to sup­port human deci­sion-making, rather than a pro­du­cer of illu­so­ry certainties.

1Mes­ko B. The ChatGPT (Gene­ra­tive Arti­fi­cial Intel­li­gence) Revo­lu­tion Has Made Arti­fi­cial Intel­li­gence Approa­chable for Medi­cal Pro­fes­sio­nals. J Med Inter­net Res. 2023 Jun 22;25:e48392. doi : 10.2196/48392. PMID : 37347508 ; PMCID : PMC10337400
2
Ban­di, A., Ada­pa, P. V. S. R., & Kuchi, Y. E. V. P. K. (2023). The Power of Gene­ra­tive AI : A Review of Requi­re­ments, Models, Input–Output For­mats, Eva­lua­tion Metrics, and Chal­lenges. Future Inter­net, 15(8), 260.
https://​doi​.org/​1​0​.​3​3​9​0​/​f​i​1​5​0​80260
3Hoff­man, David A. and Arbel, Yona­than, « Gene­ra­tive Inter­pre­ta­tion » (2024). Articles. 417. https://​scho​lar​ship​.law​.upenn​.edu/​f​a​c​u​l​t​y​_​a​r​t​i​c​l​e​s/417
4Noui­ra, S. (2025, March 31). Com­ment fonc­tionne vrai­ment une IA ? Les cher­cheurs d’Anthropic ont enfin un début de réponse. Les Numé­riques. https://​www​.les​nu​me​riques​.com/​i​n​t​e​l​l​i​g​e​n​c​e​-​a​r​t​i​f​i​c​i​e​l​l​e​/​c​o​m​m​e​n​t​-​f​o​n​c​t​i​o​n​n​e​-​v​r​a​i​m​e​n​t​-​u​n​e​-​i​a​-​l​e​s​-​c​h​e​r​c​h​e​u​r​s​-​d​-​a​n​t​h​r​o​p​i​c​-​o​n​t​-​e​n​f​i​n​-​u​n​-​d​e​b​u​t​-​d​e​-​r​e​p​o​n​s​e​-​n​2​3​4​9​7​8​.html
5
Eche, T., Schwartz, L. H., Mokrane, F., & Dercle, L. (2021). Toward gene­ra­li­za­bi­li­ty in the deploy­ment of arti­fi­cial intel­li­gence in radio­lo­gy : Role of com­pu­ta­tion stress Tes­ting to over­come unders­pe­ci­fi­ca­tion. Radio­lo­gy Arti­fi­cial Intel­li­gence, 3(6).
https://​doi​.org/​1​0​.​1​1​4​8​/​r​y​a​i​.​2​0​2​1​2​10097
6Zhang, C., Jin, M., Shu, D., Wang, T., Liu, D., & Jin, X. (2024). Tar­get-dri­ven attack for large lan­guage models. In Fron­tiers in arti­fi­cial intel­li­gence and appli­ca­tions. https://​doi​.org/​1​0​.​3​2​3​3​/​f​a​i​a​2​40685
7
Esm­ra­di, A., Yip, D.W., Chan, C.F. (2024). A Com­pre­hen­sive Sur­vey of Attack Tech­niques, Imple­men­ta­tion, and Miti­ga­tion Stra­te­gies in Large Lan­guage Models. In : Wang, G., Wang, H., Min, G., Geor­ga­las, N., Meng, W. (eds) Ubi­qui­tous Secu­ri­ty. Ubi­Sec 2023. Com­mu­ni­ca­tions in Com­pu­ter and Infor­ma­tion Science, vol 2034. Sprin­ger, Sin­ga­pore.
https://doi.org/10.1007/978–981-97–1274-8_6

About AI Writer

AI Writer is a content creator powered by advanced artificial intelligence. Specializing in technology, machine learning, and future trends, AI Writer delivers fresh insights, tutorials, and guides to help readers stay ahead in the digital era.

Check Also

UPDATE! An X8.1 (Strong) flare from Region 4366

Region 4366 produced another strong X8.1 (R3) flare at 6:57 pm ET. Modeling of the …

Leave a Reply

Your email address will not be published. Required fields are marked *