Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
And the questions are to good to be real ? I don’t think Senator’s know anythin…
ytr_UgwBE7LBi…
G
we, humanity in the west, are so, so lost. literally a product is brought to mar…
ytc_UgypJ-vuN…
G
This is a good example of how automation makes people lose all their sense of au…
ytc_Ugzy63FMo…
G
ChatGP is woke also. People need to understand that when they are letting AI hel…
ytc_Ugw6ExrmS…
G
AI Bros remind me of a goofy disney supervillain just going off by how they conv…
ytc_UgzK_bCHU…
G
Ok. I'm a fan of ai technology , and have been for a while but this was a little…
ytc_Ugy978yUP…
G
I hope this private detective is able to get something done where the law failed…
ytc_UgwVl2uog…
G
If AI goes physical and the masses can have AI, it can automate the resources to…
ytc_Ugyw_n5H-…
Comment
I just asked ChatGPT the following: Should humans fear artificial intelligence?
This is ChatGPT's answer: The question of whether humans should fear artificial intelligence (AI) is complex and depends on various factors. AI has the potential to bring about numerous benefits and advancements in many fields, such as healthcare, transportation, education, and scientific research. However, there are also valid concerns and risks associated with AI that should be addressed.
Here are some key considerations:
Benefits of AI: AI has the capacity to enhance human capabilities and improve efficiency in various industries. It can assist with tasks that are time-consuming, dangerous, or require vast amounts of data processing.
Ethical Concerns: There are ethical questions surrounding AI, particularly in areas such as privacy, bias, and the potential for AI systems to be used for harmful purposes.
Job Displacement: Automation and AI have the potential to disrupt job markets, leading to job displacement for certain industries and professions. However, new jobs and opportunities may also arise as a result of AI advancements.
Unintended Consequences: AI systems can have unintended consequences due to biases in the data they are trained on or unforeseen interactions with other systems.
Control and Regulation: Ensuring that AI systems are developed and used responsibly is crucial. There needs to be oversight and regulation to avoid harmful applications or uncontrolled proliferation of AI technology.
Superintelligence: Some experts and thinkers have raised concerns about the long-term future of AI if it were to reach superintelligence, where AI systems surpass human intelligence and capabilities. This scenario, often referred to as "artificial general intelligence" (AGI), is still speculative, but it warrants careful consideration and planning.
The key to ensuring a positive and responsible future with AI lies in how we develop, deploy, and regulate the technology. Collaborative efforts between governments, industry, researchers, and society at large can help address concerns and establish ethical guidelines for the development and use of AI.
Fear is a natural response to the unknown or potentially powerful technologies, but it should not paralyze us from exploring and utilizing AI for the benefit of humanity. Instead, a balanced approach of cautious optimism, ethical considerations, and responsible AI development can help maximize the benefits of AI while minimizing potential risks.
youtube
AI Governance
2023-07-31T22:2…
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | none |
| Reasoning | unclear |
| Policy | none |
| Emotion | indifference |
| Coded at | 2026-04-26T23:09:12.988011 |
Raw LLM Response
[
{"id":"ytc_UgzgYG4g084LCRfXNXV4AaABAg","responsibility":"government","reasoning":"consequentialist","policy":"regulate","emotion":"fear"},
{"id":"ytc_UgysDLoAUOKivpu_rRF4AaABAg","responsibility":"company","reasoning":"deontological","policy":"liability","emotion":"outrage"},
{"id":"ytc_UgyWmfJu-acslgnVbS94AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"fear"},
{"id":"ytc_Ugw8cWpdxLiti0sv2ux4AaABAg","responsibility":"developer","reasoning":"virtue","policy":"ban","emotion":"outrage"},
{"id":"ytc_UgwUh_4LCP7ZXezN6fV4AaABAg","responsibility":"user","reasoning":"consequentialist","policy":"none","emotion":"fear"},
{"id":"ytc_UgwuaaTdHueBnabn5Dx4AaABAg","responsibility":"ai_itself","reasoning":"mixed","policy":"none","emotion":"resignation"},
{"id":"ytc_Ugym462LppAW4B1akFF4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"resignation"},
{"id":"ytc_UgzTfUiI6Kt8fk6GIjJ4AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"},
{"id":"ytc_UgymJu8gmf1-COZNnZd4AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"},
{"id":"ytc_UgwA4-0agWbqiXPLe6p4AaABAg","responsibility":"developer","reasoning":"deontological","policy":"regulate","emotion":"approval"}
]