MartialTerran commited on
Commit
f87db5b
1 Parent(s): 835c7e6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +47 -0
README.md CHANGED
@@ -186,6 +186,53 @@ Here's a more understandable version of the dialog, focusing on clarity and avoi
186
 
187
  @henrismith7472: Are you all really arguing about this instead of thinking about how much this technology is going to change the world? Even if progress stopped right now, and we just focused on using what we already have, I don't think we fully grasp the impact it would have. And we have at least two more scaling laws to go, extremely efficient and powerful chips being invented, and countless accelerating technologies converging with positive feedback loops... As someone who's just started learning to code, it seems like some of you need to take a step back and look at the bigger picture.
188
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
189
  ---
190
  license: apache-2.0
191
  ---
 
186
 
187
  @henrismith7472: Are you all really arguing about this instead of thinking about how much this technology is going to change the world? Even if progress stopped right now, and we just focused on using what we already have, I don't think we fully grasp the impact it would have. And we have at least two more scaling laws to go, extremely efficient and powerful chips being invented, and countless accelerating technologies converging with positive feedback loops... As someone who's just started learning to code, it seems like some of you need to take a step back and look at the bigger picture.
188
 
189
+ Briefing Doc: Debate on the State of AI Research
190
+ Source: Online discussion thread (platform unspecified)
191
+
192
+ Date: Discussion occurred approximately two weeks prior to the document's creation
193
+
194
+ Key Participants:
195
+
196
+ @adamkadmon6339: The central figure in the debate. Writes anonymously. Expresses concerns about the current state of AI research, emphasizing a perceived lack of theoretical depth and over-reliance on scaling existing models like transformers. Claims to have deep historical knowledge of AI, including early work on backpropagation.
197
+ @ianmatejka3533: Presents counter-arguments to @adamkadmon6339, highlighting the versatility and potential of the transformer architecture and ongoing advancements in areas like reinforcement learning. Believes that transformers may be a pathway to AGI.
198
+ Other participants: Weigh in with supporting or dissenting opinions, but play a lesser role in the central argument.
199
+ Main Themes:
200
+
201
+ Theoretical Foundation vs. Empirical Success: The core tension lies in the perceived shift from strong mathematical foundations and theoretical innovation in early AI to a focus on scaling existing models and achieving impressive empirical results, particularly with large language models (LLMs) like GPT.
202
+ The Role of Transformers: The transformer architecture, while successful, is debated as being either a truly novel innovation or simply a clever recombination of existing ideas. The discussion also touches on the potential of transformers as a foundation for AGI.
203
+ Incentive Structures and Research Culture: @adamkadmon6339 argues that the current reward system in AI research favors practical applications and corporate interests over fundamental theoretical breakthroughs, contributing to a "faddish" culture.
204
+ Key Arguments:
205
+
206
+ @adamkadmon6339:
207
+ "Machine learning used to be about maths. Now it is about hardware, hype, opinions, and moving big software blocks around."
208
+ Argues that the focus on scaling large models, while yielding impressive results, lacks the deep mathematical understanding that drove earlier progress.
209
+ "Training a network for everything that happens to you is not what we do, and it is not practical in general."
210
+ Criticizes the heavy reliance on techniques like backpropagation and test-time fine-tuning without exploring more fundamentally novel approaches.
211
+ "Also: there is no money to be made from a mathematical advance unless you keep it secret...So the incentive structure of the field is screwed up and rewards those who pytorch for companies, rather than those who invest in technically deep formalisms."
212
+ Suggests that the economic incentives in AI research disincentivize theoretical work.
213
+ @ianmatejka3533:
214
+ "The transformer has been proven as a general purpose architecture across multiple domains. Researchers are now focused on more efficient learning algorithms."
215
+ Believes that the transformer architecture is a significant innovation with ample potential for further exploration and optimization.
216
+ Points to ongoing advancements in areas like reinforcement learning (RL), Monte Carlo Tree Search (MCTS), optimization algorithms, and tokenization techniques.
217
+ "The transformer isn’t the final AI architecture, but we’ve barely begun exploring its potential. There’s a wealth of low-hanging fruit yet to be studied."
218
+ Suggests that focusing on improving transformers is a more pragmatic path to AGI than seeking to "reinvent the wheel."
219
+ Possible Identity of @adamkadmon6339:
220
+
221
+ The text provides clues about @adamkadmon6339's background and expertise, suggesting they might be:
222
+
223
+ A retired or semi-retired AI/ML professor or researcher disillusioned with the current research trends.
224
+ An independent researcher with a deep understanding of AI history and a preference for theoretical work.
225
+ A senior researcher in a smaller, less prominent setting who feels free to express critical opinions anonymously.
226
+ A disgruntled former employee of a major AI lab who disagrees with the prevailing focus on scaling.
227
+ Further Research:
228
+
229
+ Stylistic analysis of @adamkadmon6339's writing could be compared to published works of potential candidates.
230
+ Investigating their online network and connections might offer additional clues to their identity.
231
+ Seeking further posts or information from @adamkadmon6339 could shed light on their background and motivations.
232
+ Conclusion:
233
+
234
+ This online discussion highlights a fundamental debate within the AI community. While the impressive results of scaled models like GPT are undeniable, concerns remain about the lack of theoretical depth and the long-term implications for the field's direction. The identity and motivations of the anonymous participant @adamkadmon6339 add an intriguing layer to the discussion.
235
+
236
  ---
237
  license: apache-2.0
238
  ---