{"id":1,"date":"2024-12-02T21:09:53","date_gmt":"2024-12-02T21:09:53","guid":{"rendered":"https:\/\/talkingtobots.net\/?p=1"},"modified":"2024-12-03T22:35:00","modified_gmt":"2024-12-03T22:35:00","slug":"hello-world","status":"publish","type":"post","link":"https:\/\/talkingtobots.net\/?p=1","title":{"rendered":"Norms of Assertion in Human-AI Communication (NIHAI)"},"content":{"rendered":"\n<p>As misinformation, fake news, and conspiracy theories circulate ever more freely, trust in media, science, and government is eroding. This challenge will intensify further, as we increasingly communicate with, and by aid of, large language models.&nbsp;<\/p>\n\n\n\n<p>Our three-year research project seeks to devise principles for responsible LLM communication: Philosophically and empirically informed rules, that specify what LLMs should and should not say. To do so, we aim to understand what people expect in conversations with AI, how they respond when these expectations aren\u2019t met, and whether their expectations and reactions differ across languages and cultures. Subsequently, we\u2019ll propose guidelines for designing AI systems that communicate responsibly and transparently. We will also test the envisioned novel guidelines with industry partners who provide language-based AI applications.&nbsp;&nbsp;<\/p>\n","protected":false},"excerpt":{"rendered":"As misinformation, fake news, and conspiracy theories circulate ever more freely, trust in media, science, and government is&hellip;\n","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[9],"tags":[],"class_list":{"0":"post-1","1":"post","2":"type-post","3":"status-publish","4":"format-standard","6":"category-blog","7":"cs-entry","8":"cs-video-wrap"},"_links":{"self":[{"href":"https:\/\/talkingtobots.net\/index.php?rest_route=\/wp\/v2\/posts\/1","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/talkingtobots.net\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/talkingtobots.net\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/talkingtobots.net\/index.php?rest_route=\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/talkingtobots.net\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=1"}],"version-history":[{"count":1,"href":"https:\/\/talkingtobots.net\/index.php?rest_route=\/wp\/v2\/posts\/1\/revisions"}],"predecessor-version":[{"id":55,"href":"https:\/\/talkingtobots.net\/index.php?rest_route=\/wp\/v2\/posts\/1\/revisions\/55"}],"wp:attachment":[{"href":"https:\/\/talkingtobots.net\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=1"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/talkingtobots.net\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=1"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/talkingtobots.net\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=1"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}