_
RobertoLofaro.com - Knowledge Portal - human-generated content
Change, with and without technology - human, AI, scraping readers welcome
for updates on publications, follow: on Instagram, Twitter, Patreon, YouTube, Kaggle metadata


_

You are now here: AI Ethics Primer - search within the bibliography - version 0.4 of 2023-12-13 > (tag cloud) >tag_selected: glaese


Currently searching for:

if you need more than one keyword, modify and separate by underscore _
the list of search keywords can be up to 50 characters long


if you modify the keywords, press enter within the field to confirm the new search key

Tag: glaese

Bibliography items where occurs: 21
The AI Index 2022 Annual Report / 2205.03468 / ISBN:https://doi.org/10.48550/arXiv.2205.03468 / Published by ArXiv / Version released on 2022-05-02 / on (web) Publishing site


Disentangling Perceptions of Offensiveness: Cultural and Moral Correlates / 2312.06861 / ISBN:https://doi.org/10.48550/arXiv.2312.06861 / Published by ArXiv / Version released on 2023-12-11 / on (web) Publishing site


AI Alignment: A Comprehensive Survey / 2310.19852 / ISBN:https://doi.org/10.48550/arXiv.2310.19852 / Published by ArXiv / Version released on 2025-04-04 / on (web) Publishing site


From human-centered to social-centered artificial intelligence: Assessing ChatGPT's impact through disruptive events / 2306.00227 / ISBN:https://doi.org/10.48550/arXiv.2306.00227 / Published by ArXiv / Version released on 2024-10-25 / on (web) Publishing site


Do LLMs Have Political Correctness? Analyzing Ethical Biases and Jailbreak Vulnerabilities in AI Systems / 2410.13334 / ISBN:https://doi.org/10.48550/arXiv.2410.13334 / Published by ArXiv / Version released on 2024-10-23 / on (web) Publishing site


Jailbreaking and Mitigation of Vulnerabilities in Large Language Models / 2410.15236 / ISBN:https://doi.org/10.48550/arXiv.2410.15236 / Published by ArXiv / Version released on 2025-11-25 / on (web) Publishing site


Hybrid Approaches for Moral Value Alignment in AI Agents: a Manifesto / 2312.01818 / ISBN:https://doi.org/10.48550/arXiv.2312.01818 / Published by ArXiv / Version released on 2025-01-16 / on (web) Publishing site


On the Trustworthiness of Generative Foundation Models: Guideline, Assessment, and Perspective / 2502.14296 / ISBN:https://doi.org/10.48550/arXiv.2502.14296 / Published by ArXiv / Version released on 2025-09-30 / on (web) Publishing site


DarkBench: Benchmarking Dark Patterns in Large Language Models / 2503.10728 / ISBN:https://doi.org/10.48550/arXiv.2503.10728 / Published by ArXiv / Version released on 2025-03-13 / on (web) Publishing site


Who is Responsible When AI Fails? Mapping Causes, Entities, and Consequences of AI Privacy and Ethical Incidents / 2504.01029 / ISBN:https://doi.org/10.48550/arXiv.2504.01029 / Published by ArXiv / Version released on 2025-09-18 / on (web) Publishing site


From Automation to Autonomy: A Survey on Large Language Models in Scientific Discovery / 2505.13259 / ISBN:https://doi.org/10.48550/arXiv.2505.13259 / Published by ArXiv / Version released on 2025-09-17 / on (web) Publishing site


A Comprehensive Survey of Deep Research: Systems, Methodologies, and Applications / 2506.12594 / ISBN:https://doi.org/10.48550/arXiv.2506.12594 / Published by ArXiv / Version released on 2025-06-14 / on (web) Publishing site


When Large Language Models Meet Law: Dual-Lens Taxonomy, Technical Advances, and Ethical Governance / 2507.07748 / ISBN:https://doi.org/10.48550/arXiv.2507.07748 / Published by ArXiv / Version released on 2025-07-10 / on (web) Publishing site


A Comprehensive Survey of Self-Evolving AI Agents: A New Paradigm Bridging Foundation Models and Lifelong Agentic Systems / 2508.07407 / ISBN:https://doi.org/10.48550/arXiv.2508.07407 / Published by ArXiv / Version released on 2025-08-31 / on (web) Publishing site


Between a Rock and a Hard Place: Exploiting Ethical Reasoning to Jailbreak LLMs / 2509.05367 / ISBN:https://doi.org/10.48550/arXiv.2509.05367 / Published by ArXiv / Version released on 2025-09-12 / on (web) Publishing site


Psychometric Personality Shaping Modulates Capabilities and Safety in Language Models / 2509.16332 / ISBN:https://doi.org/10.48550/arXiv.2509.16332 / Published by ArXiv / Version released on 2025-09-19 / on (web) Publishing site


Human-aligned AI Model Cards with Weighted Hierarchy Architecture / 2510.06989 / ISBN:https://doi.org/10.48550/arXiv.2510.06989 / Published by ArXiv / Version released on 2025-10-08 / on (web) Publishing site


Cultural Dimensions of Artificial Intelligence Adoption: Empirical Insights for Wave 1 from a Multinational Longitudinal Pilot Study / 2510.19743 / ISBN:https://doi.org/10.48550/arXiv.2510.19743 / Published by ArXiv / Version released on 2025-10-22 / on (web) Publishing site


Diverse Human Value Alignment for Large Language Models via Ethical Reasoning / 2511.00379 / ISBN:https://doi.org/10.48550/arXiv.2511.00379 / Published by ArXiv / Version released on 2025-11-01 / on (web) Publishing site


Morality in AI. A plea to embed morality in LLM architectures and frameworks / 2511.20689 / ISBN:https://doi.org/10.48550/arXiv.2511.20689 / Published by ArXiv / Version released on 2025-11-21 / on (web) Publishing site


Legal Alignment for Safe and Ethical AI / 2601.04175 / ISBN:https://doi.org/10.48550/arXiv.2601.04175 / Published by ArXiv / Version released on 2026-01-07 / on (web) Publishing site