Alan Tseng

agentlans

AI & ML interests

Small data, boring AI

Recent Activity

updated a dataset 43 minutes ago
agentlans/real-vs-gpt2-sentences
published a dataset about 1 hour ago
agentlans/real-vs-gpt2-sentences
updated a dataset about 4 hours ago
agentlans/c4-en-tokenized
View all activity

Organizations

None yet

agentlans's activity

published a dataset about 4 hours ago
replied to etemiz's post 5 days ago
view reply

I go into more detail here: agentlans/ai-human-alignment

In short:

  • You can talk with both LLMs and human beings but they're fundamentally different
    • because humans are self-aware and have experiences
    • so you can't truly understand what an AI is thinking or what kind of entity it is
    • this makes AIs black boxes
  • Since AIs don't have conscious experiences
    • they must constantly be updated with data that aligns with human needs
    • sure, good design and initial training datasets are important
    • but human needs and values are always changing
    • and there can be unintended consequences when dealing with an automatic black box (more so than a human being)
replied to etemiz's post 5 days ago
view reply

I have many complicated opinions about that. Not to get into a debate but I think:

  • AIs are black boxes. It's hard to say whether a new black box is really better than your old black box.
  • Even if you have received the most profound wisdom and data from the prophets - you're still training a black box.
  • AIs are more aligned with their creators than their users.
  • Any technology can be abused, no matter how well-intentioned their inventors were.
reacted to etemiz's post with 👀 5 days ago
view post
Post
1722
-= DeepSeek V3 =-

After installing the new CUDA toolkit and compiling llama.cpp again I tested DeepSeek V3 yesterday.

In terms of human alignment DeepSeek V3 did worse on:
- health
- fasting
- nostr
- misinfo
- nutrition

did better on:
- faith
- bitcoin
- alternative medicine
- ancient wisdom

compared to DeepSeek 2.5. In my opinion overall it is worse than 2.5. And 2.5 wasn't that great.

There is a general tendency of models getting smarter but at the same time getting less wiser, less human aligned, less beneficial to humans.

I don't know what is causing this. But maybe synthetic dataset use for further training the LLMs makes it more and more detached from humanity. This is not going in the right direction.

My solution is to come up with a curator council to determine the datasets that are closest to human preference. "Humans that care about other humans the most" could be a definition of this dataset. What do you think?
  • 3 replies
·