跳至主要内容

2025-04-03-top

  • 精選方式: TOP
  • 時間範圍: DAY

討論重點

以下是30篇文章的條列式重點總結,並附上對應的錨點連結與逐條細節:


1. I blew $417 on Claude Code to build a word game

  1. 開發成果
    • 成功用Claude開發文字遊戲LetterLinks,含棋盤設計、計分系統等功能。
  2. AI優點
    • 理解模糊指令(如「棋盤變紫色」)。
    • 快速定位錯誤代碼。
  3. AI缺點
    • 上下文限制導致後期代碼記憶困難。
    • 修正常引入新錯誤。
    • 成本飆升(簡單修改需重新理解代碼)。
  4. 經濟性
    • 成本低於真人開發,但心力消耗大。

2. Sonnet Computer Use in very Underrated

  1. 技術復興
    • Sonnet技術被低估,但潛力大。
  2. 應用案例
    • Apply HeroVercel整合瀏覽器代理。
    • BrowserUse開源商業化成功。
  3. 未來趨勢
    • 預測「網頁+電腦代理」應用爆發。

3. I regret buying claude for 1 year

  1. 版本批評
    • Claude 3.7被評為「fucking shitty」。
  2. 用戶挫折
    • 更新後體驗惡化,引發極端情緒(如自傷傾向)。

  1. 工具定位
    • 免費程式碼分析工具Serena,媲美付費方案。
  2. 技術創新
    • 使用語言伺服器解析代碼,支援MCP伺服器。
  3. 部署選項
    • 可搭配Gemini(需Google Cloud API金鑰)。

5. Migrated to Gemini. Sheeesh, the grass is greener.

  1. Claude缺點
    • 上下文長度限制、API過載錯誤。
  2. Gemini優勢
    • 100萬token上下文、更快代碼生成、無訊息限制。
  3. 應用成果
    • 順利開發本地化聊天機器人(WolfAI)。

6. What the hell happened?

  1. 投訴指南
    • 需標註環境類別、提供詳細輸入/輸出。
  2. 普遍問題
    • 多名用戶回報系統未更新(附官方狀態頁面)。

7. Please be candid; did I just pay $220 for a year of this screensaver?

  • 圖片內容無法分析,推測為迷因或諷刺Claude訂閱價值。

8. Thanks then! Take care...

  1. 技術故障
    • 功能突然失效、無故觸發長度限制錯誤。
  2. 用戶協作
    • 釐清問題根源(帳號代幣耗盡或系統端問題)。

9. Claude Sonnet is the undisputed champion of OCR

  1. 性能排名
    • Claude > 開源模型(Qwen、Mistral) > GPT-4o。
  2. 意外發現
    • OpenAI與Anthropic表現差距超出預期。

10. Claude 3.7 Sonnet is still the best LLM for frontend development

  1. LLM局限
    • 僅生成靜態模板,缺乏JS邏輯處理。
  2. 模型比較
    • Claude 3.7代碼格式不穩,Gemini 2.5 Pro更強但輸出混亂。

(因篇幅限制,以下簡化條列標題,完整細節請參照錨點連結)

11-30. 快速摘要

  • #11 從Claude轉向Gemini的情感糾結與效能比較。
  • #12 Claude Code效能佳但成本過高,盼降價。
  • #13 遊戲開發協作與發布過程的非正式討論。
  • #14 質疑Claude摘要壓縮機制犧牲上下文品質。

文章核心重點

以下是根據每篇文章標題生成的一句話摘要(條列式輸出):

  1. I blew $417 on Claude Code to build a word game. Here's the brutal truth.
    作者花費417美元用Claude開發文字遊戲,總結AI協作的高效率與成本失控的雙面性。

  2. Sonnet Computer Use in very Underrated
    分析Sonnet電腦技術被低估的現狀,預測其將因AI整合與開源商業化而復興。

  3. I regret buying claude for 1 year. It's so shit now
    用戶強烈批評Claude 3.7版本效能退化,甚至引發心理負面影響。

  4. Fully Featured AI Coding Agent as MCP Server
    介紹免費開源程式碼分析工具Serena,強調其語言伺服器技術與MCP整合優勢。

  5. Migrated to Gemini. Sheeesh, the grass is greener.
    作者轉用Gemini 2.5 Pro後,盛讚其百萬token上下文與穩定API勝過Claude。

  6. What the hell happened?
    指南說明如何有效向Claude提交問題,並反映多用戶遭遇系統異常的普遍性。

  7. Please be candid; did I just pay $220 for a year of this screensaver, but only at Anthropic's website?
    (無法解析圖片內容,推測為對Claude訂閱價值的質疑)

  8. Thanks then! Take care...
    用戶回報Claude技術故障(如無故觸發長度限制),尋求社群協作解決。

  9. Claude Sonnet is the undisputed champion of OCR
    實測顯示Claude在OCR任務表現超越GPT-4o與開源模型,引發對OpenAI技術落後的驚訝。

  10. Claude 3.7 Sonnet is still the best LLM (by far) for frontend development
    辯論Claude 3.7在前端開發的優劣,批評其基礎功能退化但仍有特定場景優勢。

  11. Pour one out for my Claude subscription... It's not you, it's Gemini (and my PhD).
    博士生因Gemini效能提升而取消Claude訂閱,流露對舊工具的情感矛盾。

  12. Claude Code was prohibitively expensive for me
    作者肯定Claude Code的編碼能力,但批評每小時21.75美元定價對自由工作者不划算。

  13. Claude created this game prototype for an idea I had in a single attempt. Now I've turned it into a full fledged game
    對話片段顯示用戶利用Claude快速開發遊戲原型並最終完成作品。

  14. Claude's context has been nerfed?
    質疑Claude實際上下文窗口可能透過摘要壓縮偷工減料,導致長對話內容遺失。

  15. Anthropic, come get your boy. Cursor + Claude 3.7 MAX
    規範要求提交AI效能證明時需附截圖與完整提示詞,否則刪文。

  16. Now we talking INTELLIGENCE EXPLOSION | Claude 3.5 cracked of benchmark!
    實驗顯示Claude 3.5在自主決策任務表現優於OpenAI,但提示設計影響結果。

  17. What Happens When You Tell an LLM It Has an iPhone Next to It?
    探討提示工程中「物件提示」對AI行為的影響,強調控制組實驗的必要性。

  18. The Impact of Generative AI on Critical Thinking - Research Paper
    研究指出GenAI重塑人類認知流程,可能削弱自主批判性思維但強化監督角色。

  19. Claude 3.7 Sonnet extended thinking is a waste of time (for deep coding) since Monday
    用戶抱怨Claude近期出現重複錯誤,懷疑系統更新導致效能異常。

  20. Dropped a new tutorial for Agentic pattern + AI SDK
    創作者發布Agentic模式教學影片並誠摯請求觀眾反饋。

  21. This conversation reached i``` maximum length
    Pro用戶抱怨服務突現過早觸發使用限制,與過去穩定體驗形成反差。

  22. Anthropic is giving free API credi for university studen
    提供學生開發者申請Anthropic免費API資源的聯繫表單連結。

  23. I'm not having issues?
    分析Claude編碼問題多源於用戶過度依賴AI或模糊提示,非工具本身缺陷。

  24. Claude Degradation Regionally Related?
    推測Claude效能問題可能與區域伺服器或網路基礎

目錄

  • [1. I blew $417 on Claude Code to build a word game. Here's the brutal truth.](#1-``` i-blew-417-on-claude-code-to-build-a-word-ga)
  • [2. Sonnet Computer Use in very Underrated](#2-``` sonnet-computer-use-in-very-underrated
- [3. ```
I regret buying claude for 1 year. It's so shit now
```](#3-```
i-regret-buying-claude-for-1-year-it-s-so-sh)
- [4. ```
Fully Featured AI Coding Agent as MCP Server
```](#4-```
fully-featured-ai-coding-agent-as-mcp-server)
- [5. ```
Migrated to Gemini. Sheeesh, the grass is greener.
```](#5-```
migrated-to-gemini-sheeesh-the-grass-is-gree)
- [6. ```
What the hell happened?
```](#6-```
what-the-hell-happened-
```)
- [7. ```
Please be candid; did I just pay $220 for a year of this screensaver, but only at Anthropic's website?
```](#7-```
please-be-candid;-did-i-just-pay-220-for-a-y)
- [8. ```
Thanks then! Take care...
```](#8-```
thanks-then-take-care-
```)
- [9. ```
Claude Sonnet is the undisputed champion of OCR
```](#9-```
claude-sonnet-is-the-undisputed-champion-of-)
- [10. ```
Claude 3.7 Sonnet is still the best LLM (by far) for frontend development
```](#10-```
claude-3-7-sonnet-is-still-the-best-llm-by-)
- [11. ```
Pour one out for my Claude subscription... It's not you, it's Gemini (and my PhD).
```](#11-```
pour-one-out-for-my-claude-subscription-it-)
- [12. ```
Claude Code was prohibitively expensive for me
```](#12-```
claude-code-was-prohibitively-expensive-for)
- [13. ```
Claude created this game prototype for an idea I had in a single attempt. Now I've turned it into a full fledged game
```](#13-```
claude-created-this-game-prototype-for-an-i)
- [14. ```
Claude's context has been nerfed?
```](#14-```
claude-s-context-has-been-nerfed-
```)
- [15. ```
Anthropic, come get your boy. Cursor + Claude 3.7 MAX
```](#15-```
anthropic-come-get-your-boy-cursor-claude-3)
- [16. ```
Now we talking INTELLIGENCE EXPLOSION | Claude 3.5 cracked of benchmark!
```](#16-```
now-we-talking-intelligence-explosion-|-cla)
- [17. ```
What Happens When You Tell an LLM It Has an iPhone Next to It?
```](#17-```
what-happens-when-you-tell-an-llm-it-has-an)
- [18. ```
The Impact of Generative AI on Critical Thinking - Research Paper
```](#18-```
the-impact-of-generative-ai-on-critical-thi)
- [19. ```
Claude 3.7 Sonnet extended thinking is a waste of time (for deep coding) since Monday
```](#19-```
claude-3-7-sonnet-extended-thinking-is-a-wa)
- [20. ```
Dropped a new tutorial for Agentic pattern + AI SDK
```](#20-```
dropped-a-new-tutorial-for-agentic-pattern-)
- [21. ```
This conversation reached i``` maximum length
```](#21-```
this-conversation-reached-i```-maximum-leng)
- [22. Anthropic is giving free API credi``` for university studen```](#22-anthropic-is-giving-free-api-credi```-for-unive)
- [23. ```
I'm not having issues?
```](#23-```
i-m-not-having-issues-
```)
- [24. ```
Claude Degradation Regionally Related?
```](#24-```
claude-degradation-regionally-related-
```)
- [25. ```
Claude Suddenly Cant Summarize Previous Cha``` by UUID?
```](#25-```
claude-suddenly-cant-summarize-previous-cha)
- [26. ```
is it just me or has clause 3.5 gone bonkers today?
```](#26-```
is-it-just-me-or-has-clause-3-5-gone-bonker)
- [27. ```
Notion MCP tutorial
```](#27-```
notion-mcp-tutorial
```)
- [28. ```
Best way to inject a prior chat history seamlessly into a current chat?
```](#28-```
best-way-to-inject-a-prior-chat-history-sea)
- [29. ```
Worked on a lofi platform
```](#29-```
worked-on-a-lofi-platform
```)
- [30. ```
Did Claude get smarter again?
```](#30-```
did-claude-get-smarter-again-
```)

---

## 1. ```
I blew $417 on Claude Code to build a word game. Here's the brutal truth.
``` {#1-```
i-blew-417-on-claude-code-to-build-a-word-ga}

這篇文章的核心討論主題是:**作者使用AI助手(Claude)開發一款類似Scrabble的文字遊戲(LetterLinks)的經驗總結與反思**。

主要分為以下幾個重點:
1. **開發成果**:成功利用AI快速實現遊戲核心功能(如棋盤設計、計分系統、每日挑戰等),並對最終成果感到滿意。
2. **AI協作的優點**:
- 能理解模糊指令並生成有效代碼(如直接回應「讓棋盤變成紫色」)。
- 部分除錯效率高(如快速定位錯誤代碼位置)。
3. **AI協作的缺點與挑戰**:
- **上下文限制**:代碼量增大後,AI難以記住前期討論內容。
- **不可靠的修正**:常出現「假性解決問題」或引入新錯誤。
- **成本失控**:後期簡單修改因需重新理解代碼而費用飆升。
- **缺乏自主測試**:需人工反覆驗證每項修改。
4. **經濟性評估**:認為AI成本仍低於雇用真人開發者,但過程耗費大量心力。
5. **未來改進方向**:模組化開發、預算管控、降低預期等。

整體而言,文章是對「AI輔助開發」實際體驗的批判性回顧,強調其效率與經濟優勢,同時揭露技術局限性與管理痛點。

- **Reddit 連結**: [https://reddit.com/r/ClaudeAI/comments/1jpddbf/i_blew_417_on_claude_code_to_build_a_word_game/](https://reddit.com/r/ClaudeAI/comments/1jpddbf/i_blew_417_on_claude_code_to_build_a_word_game/)
- **外部連結**: [https://www.reddit.com/r/ClaudeAI/comments/1jpddbf/i_blew_417_on_claude_code_to_build_a_word_game/](https://www.reddit.com/r/ClaudeAI/comments/1jpddbf/i_blew_417_on_claude_code_to_build_a_word_game/)
- **發布時間**: 2025-04-02 10:01:59

### 內容

Alright, so a few weeks ago ago I had this idea for a Scrabble-style game and thought "why not try one of these fancy AI coding assistan?" Fast forward through a sh\*t ton of prompting, $417 in Claude credi, and enough coffee to kill a small horse, I've finally got a working game called LetterLinks: https://playletterlinks.com/

The actual game (if you care)

It's basically my take on Scrabble/Wordle with daily challenges:

- Place letter tiles on a board

  - Form words, get poin```

- Daily themes and bonus challenges

- Leaderboards to flex on strangers

The Good Par``` (there were some)

Actually nailed the implementation

I literally started with "make me a scrabble-like game" and somehow Claude understood what I meant. No mockups, no wireframes, just me saying "make the board purple" or "I need a timer" and it spitting out working code. Not gonna lie, that part was pretty sick.

Once I described a feature I wanted - like skill levels that show progress - Claude would run with it.

Ultimately I think the finished result is pretty slick, and while there are some bugs, I'm proud of what Claude and I did together.

Debugging that didn't always completely suck

When stuff broke (which was constant), conversations often went like:

Me: "The orange multiplier badges are showing the wrong number"

Claude: dumps exact code location and fix

This happened often enough to make me not throw my laptop out the window.

The Bad Par``` (oh boy)

Context window is a giant middle finger

Once the codebase hit about 15K lines, Claude basically became that friend who keeps asking you to repeat the story you just told:

Me: "Fix the bug in the theme detection

Claude: "What theme detection?"

Me: "The one we've been working on FOR THE PAST WEEK"

I had to use the /claude compact feature more and more frequently.

The "I found it!" BS

Most irritating phrase ever:

Claude: "I found the issue! It's definitely this line right here."

implemen``` fix

bug still exis```

Claude: "Ah, I see the REAL issue now..."

Rinse and repeat until you're questioning your life choices. Bonus poin``` when Claude confidently "fixes" something and introduces three new bugs.

Cost spiral is real

What really pissed me off was how the cost scaled:

- First week: Built most of the game logic for ~$100

- Last week: One stupid animation fix cost me $20 because Claude needed to re-learn the entire codebase

The biggest "I'm never doing this again but probably will" part

Testing? What testing?

Every. Single. Change. Had to be manually tested by me. Claude can write code all day but can't click a f***ing button to see if it works.

This turned into:

  1. Claude writes code

  2. I test

  3. I report issues

  4. Claude apologizes and tries again

  5. Repeat until I'm considering a career change

Worth it?

For $417? Honestly, yeah, kinda. A decent freelancer would have charged me $2-3K minimum. Also I plan to use this in my business, so it's company money, not mine. But it wasn't the magical experience they sell in the ads.

Think of Claude as that junior dev who sometimes has brilliant ideas but also needs constant supervision and occasionally se``` your project on fire.

Next time I'll:

  1. Split everything into tiny modules from day one

  2. Keep a separate doc with all the architecture decisions

  3. Set a hard budget per feature

  4. Lower my expectations substantially

Anyone else blow their money on AI coding? Did you have better luck, or am I just doing it wrong?


### 討論

**評論 1**:

Gemini 2.5 Pro experimental will solve your context woes


**評論 2**:

Interesting post. I have been experimenting with Claude Code at work. We have a relatively big codebase (a web application that has been developed by more than a dozen developers over the last 6 years). But it's also written in a very professional way, using all the best practices, standards, etc. Claude Code does very well in such an environment. I it fix two dozen papercut bugs (usually smaller issues like icons not displaying correcctly, time displaying in UTC instead of the browser timezone, dropdown menu not showing some items correctly, etc) and it fixed them all pretty much spot-on. Most of them cost less than $1 to fix with Claude Code. I should also mention that we have a professional QA team that logs ticke very well with detailed steps to reproduce, expected resul, actual resul```, etc, which makes it easier for Calude Code to find the right code and fix it well.

I have it currently monitoring our JIRA ticke, and when a ticket with "papercut" label shows up, Claude Code tries to fix it and submi the fix as a PR with all the explanation on how it was thinking. This makes it quite easy for developers to review. It also helps that our CI/CD will run a battery of tes on the proposed code before it ge to a developer for review.

So from my experience, if you have well-written codebase, and you explain in detail what you want it to do, it does well.


**評論 3**:

One thing I forgot to add - feature creep. I would often ask Claude for a relatively simple addition, but it would take it upon ielf to add 5 different things. Sometimes they were actually good ideas, but often they weren't and I'd waste time and money by telling to change things back. I also suspect that, although the game works pretty well afaict, my actual codebase is a bit of a mess with lo of redundancies and abandoned sections (I don't know for sure because I'm not a coder - but just from skimming over the source code it seems like some functions have been repeated in multiple par```).


**評論 4**:

My secret trick is, I use git2text on the codebase. I put everything in Google AI Studio (G pro 2.5), it comment on the code, and have it generate specific to-do lis``` that I paste back in the claude code. Works fantastic.


**評論 5**:

Pretty fun game, I dig it.


---

## 2. ```
Sonnet Computer Use in very Underrated
``` {#2-```
sonnet-computer-use-in-very-underrated
```}

這篇文章的核心討論主題是 **「Sonnet 電腦技術的潛力與近期發展」**,並強調以下重點:

1. **Sonnet 技術的復興**:作者認為這項技術目前被低估,但即將迎來重大回歸。
2. **實際應用案例**:
- 熱門應用(如 *Apply Hero*)的興起。
- *Vercel* 的 AI SDK 整合了瀏覽器代理(web agen```),顯示技術整合趨勢。
- *BrowserUse* 透過開源授權獲利,反映商業潛力。
- 爆紅演示 *Manus* 主要基於 Sonnet 開發。
3. **未來展望**:預測更多結合「網頁與電腦代理」(web + computer use agen```)的應用即將爆發。

**關鍵詞**:Sonnet、電腦代理(computer use agen```)、瀏覽器技術、AI 整合、開源商業化。

- **Reddit 連結**: [https://reddit.com/r/ClaudeAI/comments/1jptm7p/sonnet_computer_use_in_very_underrated/](https://reddit.com/r/ClaudeAI/comments/1jptm7p/sonnet_computer_use_in_very_underrated/)
- **外部連結**: [https://www.reddit.com/r/ClaudeAI/comments/1jptm7p/sonnet_computer_use_in_very_underrated/](https://www.reddit.com/r/ClaudeAI/comments/1jptm7p/sonnet_computer_use_in_very_underrated/)
- **發布時間**: 2025-04-03 00:58:01

### 內容

I dove really deep into browser agen``` for the past month and sonnet computer use is very impressive. I think its currently very underrated, and going to have a huge comeback.

- Very cool computer use applications like Apply Hero are becoming very popular.

- AI SDK by Vercel (most used typescript AI sdk) just integrated web agen``` today.

- BrowserUse rumor is that they are making millions of ARR by just licensing their open source project.

- Manus (the very viral demo) is built mostly on sonnet.

I think so many more web + computer use agen``` are going to popoffverysoon.


### 討論

**評論 1**:

What kind of problems can it solve for a general public everyday PC user?


**評論 2**:

Do they have a dedicated SDK?


**評論 3**:

I do not feel like we are there yet, as much as everyone wan``` to believe.

Anthropic keeps shitting the bed every time some progress is being made, removing features, blocking usage, and its clear MCP has proven very effective in making AI agen``` but the infrastructure is not there to support it. Chat GPT supporting MCP even further proves that its going to be a new standard.

Manus has been proven to not be ready for production, and was a complete dud. I have been joined in their group since day one.


---

## 3. ```
I regret buying claude for 1 year. It's so shit now
``` {#3-```
i-regret-buying-claude-for-1-year-it-s-so-sh}

這篇文章的核心討論主題是對「Claude 3.7」版本(可能是AI模型或軟體)的強烈負面評價,使用者以情緒化的語言表達對該版本的極度不滿,甚至提到對自身心理狀態的負面影響(如「kms」為自殺相關縮寫)。

重點歸納:
1. **版本批評**:直指「Claude 3.7」功能或表現差勁(用「fucking shitty」等俚語強化情緒)。
2. **使用者挫折感**:反映版本更新可能導致體驗惡化,引發強烈負面情緒。
3. **潛在極端反應**:文中提及自我傷害的極端表述(需注意此類言論可能涉及心理健康議題)。

註:若此為真實求助內容,建議關注發文者的心理狀態並提供必要支持。

- **Reddit 連結**: [https://reddit.com/r/ClaudeAI/comments/1jpk5m3/i_regret_buying_claude_for_1_year_its_so_shit_now/](https://reddit.com/r/ClaudeAI/comments/1jpk5m3/i_regret_buying_claude_for_1_year_its_so_shit_now/)
- **外部連結**: [https://www.reddit.com/r/ClaudeAI/comments/1jpk5m3/i_regret_buying_claude_for_1_year_its_so_shit_now/](https://www.reddit.com/r/ClaudeAI/comments/1jpk5m3/i_regret_buying_claude_for_1_year_its_so_shit_now/)
- **發布時間**: 2025-04-02 17:03:19

### 內容

Cluade 3.7 is fucking shitty and is gonna make me kms


### 討論

**評論 1**:

Cancelled last month lol. So many models now, why even pay. Generate with one, bug fix with another, organize with other. Easy.


**評論 2**:

with the absolute insane pace of this industry, why would you lock in with 1 provider for a whole year?


**評論 3**:

I want opus 4.0


**評論 4**:

The cycle of AI releases:

A new Model -> Finally <insert AI> is good, i could do X

just a general thing without any detail

For the next month -> <insert AI> is revolutionary, I could do so many things

fails to explain how

After any other model of any other AI does something slightly better -> I will literally kms because <insert AI> is bad

fails to explain why

Go to step 1 for the new AI.


**評論 5**:

Dare to elaborate what you find shitty about it? I've figured that GPT 4o is better at some tasks such as rephrasing text whereas Claude 3.7 performs much better with coding and calculating (and corn).


---

## 4. ```
Fully Featured AI Coding Agent as MCP Server
``` {#4-```
fully-featured-ai-coding-agent-as-mcp-server}

這段文章的核心討論主題是:
**「介紹一款免費且功能強大的程式碼分析代理工具(Serena),並說明其技術特點與應用方式。」**

具體要點包括:
1. **工具定位**:免費提供與付費工具(如Windsurf的Cascade或Cursor的agent)同等或更優的程式碼分析能力。
2. **技術創新**:
- 使用**語言伺服器(language server)**而非傳統RAG技術來解析程式碼,提升對大型程式碼庫的理解能力。
- 支援作為**MCP伺服器**運行,可與Claude Desktop免費整合。
3. **部署選項**:
- 可搭配Gemini,但需Google Cloud API金鑰(新用戶享300美元贈金)。
4. **開源授權**:以GPL許可證釋出,強調易用性與可近性。

附帶目標是推廣其GitHub專案([Serena](https://github.com/oraios/serena))。

- **Reddit 連結**: [https://reddit.com/r/ClaudeAI/comments/1jpavtm/fully_featured_ai_coding_agent_as_mcp_server/](https://reddit.com/r/ClaudeAI/comments/1jpavtm/fully_featured_ai_coding_agent_as_mcp_server/)
- **外部連結**: [https://www.reddit.com/r/ClaudeAI/comments/1jpavtm/fully_featured_ai_coding_agent_as_mcp_server/](https://www.reddit.com/r/ClaudeAI/comments/1jpavtm/fully_featured_ai_coding_agent_as_mcp_server/)
- **發布時間**: 2025-04-02 08:00:46

### 內容

We've been working like hell on this one: a fully capable Agent, as good or better than Windsurf's Cascade or Cursor's agent - but can be used for free.

It can run as an MCP server, so you can use it for free with Claude Desktop, and it can still fully understand a code base, even a very large one. We did this by using a language server instead of RAG to analyze code.

Can also run it on Gemini, but you'll need an API key for that. With a new google cloud account you'll get 300$ as a gift that you can use on API credi```.

Check it out, super easy to run, GPL license:

https://github.com/oraios/serena


### 討論

**評論 1**:

ELI5, how is this different than Cline or Roo?


**評論 2**:

Can you describe how this is different from the bash scripting MCPs (like wcgw) that are also optimised for coding?

Those essentially run like Claude Code and are fantastic, especially on large projec``` (my experience at least). Im wondering if this is similar or does it provide additional functionality.


**評論 3**:

Im using it right now. Its like read/write on steroids! Amazing! Is there a discord group we can make to stay connected in a community?


**評論 4**:

Can we use it for php? With Gemini 2.5 pro?


**評論 5**:

Any chance of NodeJS or any other variant for programming options?


---

## 5. ```
Migrated to Gemini. Sheeesh, the grass is greener.
``` {#5-```
migrated-to-gemini-sheeesh-the-grass-is-gree}

這篇文章的核心討論主題是:**作者從Claude API轉換到Gemini 2.5 Pro的體驗比較與優點**,主要聚焦於以下幾點:

1. **對Claude API的不滿**:
- 受上下文長度限制(contextual limit)困擾
- 頻繁出現API過載錯誤(`api overload errors`)

2. **Gemini 2.5 Pro的優勢**:
- 更高的性價比(相較於每月20美元的方案)
- 支援**100萬token的上下文長度**(1M token limit)
- 更快的代碼生成與更新速度
- 無訊息限制(no limit messages)
- 順利完成本地化聊天機器人(localized chatbot)開發

3. **實際應用成果**:
- 透過RooCode工具流暢構建項目
- 快速迭代開發(trial and error過程效率提升)
- 附上個人項目鏈接(WolfAI)作為成功案例

整體而言,這是一篇**開發者對比AI API服務的實測心得**,強調Gemini在處理長上下文、穩定性和開發效率上的改進。

- **Reddit 連結**: [https://reddit.com/r/ClaudeAI/comments/1jpk827/migrated_to_gemini_sheeesh_the_grass_is_greener/](https://reddit.com/r/ClaudeAI/comments/1jpk827/migrated_to_gemini_sheeesh_the_grass_is_greener/)
- **外部連結**: [https://www.reddit.com/r/ClaudeAI/comments/1jpk827/migrated_to_gemini_sheeesh_the_grass_is_greener/](https://www.reddit.com/r/ClaudeAI/comments/1jpk827/migrated_to_gemini_sheeesh_the_grass_is_greener/)
- **發布時間**: 2025-04-02 17:08:35

### 內容

Honestly, i just got tired of the contextual limit via the Claude API. Cost wise, it was a better option than the $20 a month. BUT MAN! Switched my API to Gemini 2.5 Pro and RooCode just hums away building everything with no limit messages or api overload errors.. The 1M token limit is dope, but I've noticed that even the code updates are way faster. Built my localized chatbot with some trial and error incredibly fast.

https://wolfai.us


### 討論

**評論 1**:

The recent clamp downs on the limi``` have forced me to use other models and Gemini 2.5 pro came just in time, it's at LEAST as good as extended claude if not clearly better most of the time. I hated all the previous google models but this one is good tbh.


**評論 2**:

I've switched away from Claude (to ChatGPT 4o) because I'm just getting started with a conversation when I get warnings about chat length. I switched to 4o really out of necessity, and I've been impressed with it. I'm in one chat that's been going on for multiple days with hundreds of responses, and it's still remembering and joking about things I said at the beginning. Claude's become unusuable for me, except for quick hit and run queries. I'm paying $20/month, same as ChatGPT, so makes me wonder why I'm paying when I'm being shut out after low reasonable-usage length cha```?


**評論 3**:

No way to make Gemini get local file access?


**評論 4**:

How do you use it without limitation I’m keep hitting limi```

**評論 5**:

I literally don't believe you. where are all these Gemini shills coming from


---

## 6. ```
What the hell happened?
``` {#6-```
what-the-hell-happened-
```}

這篇文章的核心討論主題是:**用戶在向Claude AI提出投訴或問題時應遵循的指南與注意事項**,同時也反映了部分用戶遇到相似問題(如系統未更新)的共鳴。

重點包括:
1. **投訴格式建議**:
- 選擇正確的環境標籤(免費/付費網頁版、API)。
- 提供詳細資訊(如輸入提示和輸出結果)以利問題診斷。
- 理解相同條件下可能因Anthropic測試機制而產生不同結果。
- 對不滿意的輸出按「拇指向下」以反饋給開發團隊。

2. **用戶反饋**:
- 多名用戶回報遇到相同問題(如系統狀態未更新),並附上官方狀態頁面連結供參考。
- 簡短回應(如「same here」)顯示問題的普遍性。

3. **自動化說明**:
- 文末註明為自動生成訊息,並引導用戶聯繫版主以進一步協助。

整體聚焦於「如何有效提交問題」及「當前服務異常的集體回報」。

- **Reddit 連結**: [https://reddit.com/r/ClaudeAI/comments/1jpa0nb/what_the_hell_happened/](https://reddit.com/r/ClaudeAI/comments/1jpa0nb/what_the_hell_happened/)
- **外部連結**: [https://i.redd.it/cx8kwupb4bse1.jpeg](https://i.redd.it/cx8kwupb4bse1.jpeg)
- **發布時間**: 2025-04-02 07:21:08

### 內容

When making a complaint, please

  1. make sure you have chosen the correct flair for the Claude environment that you are using: i.e Web interface (FREE), Web interface (PAID), or Claude API. This information helps others understand your particular situation.

  2. try to include as much information as possible (e.g. prompt and output) so that people can understand the source of your complaint.

  3. be aware that even with the same environment and inpu```, others might have very different outcomes due to Anthropic's testing regime.

  4. be sure to thumbs down unsatisfactory Claude output on Claude.ai. Anthropic representatives tell us they monitor this data regularly.

I am a bot, and this action was performed automatically. Please contact the moderators of this subreddit if you have any questions or concerns.

Yeah same problem.

same here. No update from their status page too

https://status.anthropic.com/

Yup same for me.

Let him think.


### 討論

**評論 1**:

When making a complaint, please

  1. make sure you have chosen the correct flair for the Claude environment that you are using: i.e Web interface (FREE), Web interface (PAID), or Claude API. This information helps others understand your particular situation.
  2. try to include as much information as possible (e.g. prompt and output) so that people can understand the source of your complaint.
  3. be aware that even with the same environment and inpu```, others might have very different outcomes due to Anthropic's testing regime.
  4. be sure to thumbs down unsatisfactory Claude output on Claude.ai. Anthropic representatives tell us they monitor this data regularly.

I am a bot, and this action was performed automatically. Please contact the moderators of this subreddit if you have any questions or concerns.


**評論 2**:

Yeah same problem.


**評論 3**:

same here. No update from their status page too

https://status.anthropic.com/


**評論 4**:

Yup same for me.


**評論 5**:

Let him think.


---

## 7. ```
Please be candid; did I just pay $220 for a year of this screensaver, but only at Anthropic's website?
``` {#7-```
please-be-candid;-did-i-just-pay-220-for-a-y}

由於提供的連結是圖片格式(Reddit 預覽圖),我無法直接讀取或分析其中的文字內容。不過,可以根據常見的討論情境推測可能的討論主題:

1. **圖片內容推測**:
- 若圖片是 **迷因(Meme)或諷刺圖**,可能涉及網路文化、社會現象或時事調侃。
- 若為 **新聞截圖或數據圖表**,可能討論政治、經濟、科技等議題(例如加密貨幣、AI 發展等)。
- 若屬 **個人創作或攝影**,主題可能圍繞藝術表達或攝影技巧。

2. **常見 Reddit 討論方向**:
- 圖片標題或評論區的互動可能透露核心主題(例如:社會議題、流行文化、技術問題)。
- 若來自特定子版塊(如 `r/technology`、`r/funny`),可進一步縮小範圍。

3. **建議下一步**:
- 提供圖片的文字描述或子版塊名稱,以便更精確總結。
- 直接分享圖片中的關鍵文字內容。

如需更準確的分析,請補充更多資訊!

- **Reddit 連結**: [https://reddit.com/r/ClaudeAI/comments/1jpaadi/please_be_candid_did_i_just_pay_220_for_a_year_of/](https://reddit.com/r/ClaudeAI/comments/1jpaadi/please_be_candid_did_i_just_pay_220_for_a_year_of/)
- **外部連結**: [https://www.reddit.com/r/ClaudeAI/comments/1jpaadi/please_be_candid_did_i_just_pay_220_for_a_year_of/](https://www.reddit.com/r/ClaudeAI/comments/1jpaadi/please_be_candid_did_i_just_pay_220_for_a_year_of/)
- **發布時間**: 2025-04-02 07:33:10

### 內容

https://preview.redd.it/wkdat48b6bse1.jpg?width=1920&format=pjpg&auto=webp&s=bbbc185a4528ad74e6ec4369f451c4a458e674c6


### 討論

**評論 1**:

Starting to get kind of ridiculous how often its been down lately


**評論 2**:

You are paying so that the people at Claude can eat well . That should make you feel better.


**評論 3**:

Mate, I just renabled my subscription on Claude after ages. Installed Claude Desktop and immediately regretted. I am bettee off Windsurf or even Roo Cline or similar using Gemini/Claude API but the application and the WebUI sucks! Ran into length limi``` so many times that I had to switch back to API based coding immediately.


**評論 4**:

They struggle with thinking mode. And it adds a lot of load too.
And I'm using mostly the thinking more to use plan from o3 and Gemini now.
Anthropic need to rework the Thinking mode. But the execution remain very good.


**評論 5**:

Im working on building an entire web app from scratch with AI help. Im not a coder at all. I first started with Gemini (consumer web interface) using 2.0 flash. When that got stuck , I took my work and went to Claude Sonnet 3.7 which Im paying for monthly. It was going VERY well so much better but yes I still had to deal with chat limi and time limi like everyone but the resul``` were going great and troubleshooting as well.

THEN everyone started raving about Gemini 2.5 and that I should also use AI studio. The main problems with AI Studio: unbearable input lag after only 50-60k tokens , no project knowledge like Claude so I cant just connect it to my Github repo and a convenient sync I have to zip up the latest version and share for every new chat.

Because of the unbearable input lag with AI studio, and no suggested troubleshooting will solve it , I had to go back to Claude and just deal with i``` current shortcomings that interrupt , along with the Continue Continue Continue aside from chat length AND session length interruptions

Its a good thing Im only paying monthly but this whole Gemini 2.5 and 1M context breaking after 60k tokens is utter nonsense.

Damned if I do and damned if I dont. Im just working with Claude on more efficient update process even though I prefer it just generate entire new files for me to just replace than dig through and replace a .


---

## 8. ```
Thanks then! Take care...
``` {#8-```
thanks-then-take-care-
```}

這組對話的核心討論主題是 **用戶在使用Claude AI時遇到的技術問題與投訴建議**,主要包含以下重點:

1. **投訴指南**
- 原始貼文提供提交問題的具體建議(如標記正確環境類別、提供詳細輸入/輸出資訊),強調用戶反饋對Anthropic團隊的重要性(如點擊「不滿意」按鈕)。

2. **技術故障回報**
- 後續回覆反映多項操作異常:
* 功能突然失效("was working fine until last week")
* 無故觸發長度限制錯誤(即使開啟新對話)
* 伺服器繁忙或配額問題的錯誤提示

3. **用戶協作需求**
- 對話顯示用戶試圖釐清問題根源(如是否為帳號代幣耗盡或系統端問題),並尋求社群協助解決當前使用障礙。

整體聚焦於「Claude服務異常時的標準化問題回報流程」與「近期普遍性技術故障的社群討論」。

- **Reddit 連結**: [https://reddit.com/r/ClaudeAI/comments/1jpjhvv/thanks_then_take_care/](https://reddit.com/r/ClaudeAI/comments/1jpjhvv/thanks_then_take_care/)
- **外部連結**: [https://i.redd.it/4xobuf5gqdse1.png](https://i.redd.it/4xobuf5gqdse1.png)
- **發布時間**: 2025-04-02 16:12:02

### 內容

When making a complaint, please

  1. make sure you have chosen the correct flair for the Claude environment that you are using: i.e Web interface (FREE), Web interface (PAID), or Claude API. This information helps others understand your particular situation.

  2. try to include as much information as possible (e.g. prompt and output) so that people can understand the source of your complaint.

  3. be aware that even with the same environment and inpu```, others might have very different outcomes due to Anthropic's testing regime.

  4. be sure to thumbs down unsatisfactory Claude output on Claude.ai. Anthropic representatives tell us they monitor this data regularly.

I am a bot, and this action was performed automatically. Please contact the moderators of this subreddit if you have any questions or concerns.

Oh I thought I was the only one. Hopefully this is some bad setting as it was working fine until last week.

Yes, Cuck flaude

Can someone explain why this is showing? Your account is out of tokens or the servers are too busy?

Also getting the "Your message will exceed the length limit for this chat. Try shortening your message or starting a new conversation." message even if I start a new chat and new project. Hopefully will be fixed as stranded now.


### 討論

**評論 1**:

When making a complaint, please

  1. make sure you have chosen the correct flair for the Claude environment that you are using: i.e Web interface (FREE), Web interface (PAID), or Claude API. This information helps others understand your particular situation.
  2. try to include as much information as possible (e.g. prompt and output) so that people can understand the source of your complaint.
  3. be aware that even with the same environment and inpu```, others might have very different outcomes due to Anthropic's testing regime.
  4. be sure to thumbs down unsatisfactory Claude output on Claude.ai. Anthropic representatives tell us they monitor this data regularly.

I am a bot, and this action was performed automatically. Please contact the moderators of this subreddit if you have any questions or concerns.


**評論 2**:

Oh I thought I was the only one. Hopefully this is some bad setting as it was working fine until last week.


**評論 3**:

Yes, Cuck flaude


**評論 4**:

Can someone explain why this is showing? Your account is out of tokens or the servers are too busy?


**評論 5**:

Also getting the "Your message will exceed the length limit for this chat. Try shortening your message or starting a new conversation." message even if I start a new chat and new project. Hopefully will be fixed as stranded now.


---

## 9. ```
Claude Sonnet is the undisputed champion of OCR
``` {#9-```
claude-sonnet-is-the-undisputed-champion-of-}

這段文字的核心討論主題是:**比較不同AI模型(Claude、GPT-4o、開源模型如Qwen和Mistral)在特定使用場景下的性能表現**。作者通過測試得出結論,認為在此用例中Claude明顯優於其他模型,而GPT-4o甚至落後於部分開源競爭對手,並對OpenAI與Anthropic(Claude開發公司)之間的差距表示驚訝。此外,作者附上測試影片連結並邀請讀者反饋。

關鍵要點:
1. **性能排名**:Claude > 開源模型(Qwen、Mistral) > GPT-4o
2. **意外發現**:OpenAI(GPT-4o)與Anthropic(Claude)的表現差距超出預期。
3. **測試背景**:基於具體用例的實測結果,非全面性評比。

- **Reddit 連結**: [https://reddit.com/r/ClaudeAI/comments/1jpne7u/claude_sonnet_is_the_undisputed_champion_of_ocr/](https://reddit.com/r/ClaudeAI/comments/1jpne7u/claude_sonnet_is_the_undisputed_champion_of_ocr/)
- **外部連結**: [https://www.reddit.com/r/ClaudeAI/comments/1jpne7u/claude_sonnet_is_the_undisputed_champion_of_ocr/](https://www.reddit.com/r/ClaudeAI/comments/1jpne7u/claude_sonnet_is_the_undisputed_champion_of_ocr/)
- **發布時間**: 2025-04-02 20:31:09

### 內容

Hey all, so put a lot of time and burnt a ton of tokens testing this, so hope you all find it useful. TLDR - Claude is the clear winner here, and GPT-4o is behind even opensource competitors like Qwen and Mistral. Very surprised for the gap between openai and anthropic in this use case!

I welcome your feedback...

https://youtu.be/ZTJmjhMjlpM


### 討論

**評論 1**:

When submitting proof of performance, you must include all of the following:

  1. Screensho``` of the output you want to report
  2. The full sequence of promp``` you used that generated the output, if relevant
  3. Whether you were using the FREE web interface, PAID web interface, or the API if relevant

If you fail to do this, your post will either be removed or reassigned appropriate flair.

Please report this post to the moderators if does not include all of the above.

I am a bot, and this action was performed automatically. Please contact the moderators of this subreddit if you have any questions or concerns.


**評論 2**:

Very cool analysis. As someone who manages a lot of scientific pdfs, it would also be interesting to see strategies for managing citations as well as plo```/figures (which I guess would need to be clipped and stored).

The semantic layout of html for future LLM ingestion hadnt been on my mind but now that you highlight it, its so obviously important. Thanks!


**評論 3**:

For ocr works much better Gemini 3 27b or mistal for me


**評論 4**:

This is good to know for an OCR project I'm starting next month.

I'm creating an open source PDF converter that adds accessiblity features to an existing non-accessible PDF. My strategy is to reverse-engineer a PDF to a source document format, make it accessible, save as a new PDF, and eval it for readable formatting.

This video is doing something similar and will be very useful to me.


**評論 5**:

Its case by case, in my experience 4o had beaten all other options at extracting and formatting text from an old edition of a thesaurus with complex and varied page layou```.


---

## 10. ```
Claude 3.7 Sonnet is still the best LLM (by far) for frontend development
``` {#10-```
claude-3-7-sonnet-is-still-the-best-llm-by-}

這四段討論的核心主題可以總結為:**對當前大型語言模型(LLM)在編程與前端開發能力上的評估與比較**,具體聚焦於以下幾個面向:

1. **LLM在前端開發中的局限性**
- 指出LLM僅能生成表面靜態的HTML/CSS模板,但缺乏處理真正前端開發核心(如JavaScript邏輯、API整合、效能優化等)的能力。
- 強調若想全面評估LLM,需測試更複雜的前端開發層面,而非僅關注視覺外觀或SEO。

2. **模型間的效能差異與用戶體驗**
- 對Claude 3.7的批評:基本代碼格式(如縮進)都無法穩定輸出,顯示其編碼能力退化。
- 對Gemini 2.5 Pro的評價:能力超越Claude且成本更低,但輸出結果被描述為「混亂」(chaotic),暗示其雖強大但穩定性不足。

3. **開發者對不同模型的偏好與實用性比較**
- 用戶從Claude Sonnet的支持者轉向Gemini,反映模型競爭中性能與性價比的現實取捨。
- 間接點出LLM在「解決複雜問題」與「基礎任務可靠性」之間可能存在的能力不平衡。

**整體核心**:討論LLM在技術開發領域的實際應用邊界,並透過具體案例(前端開發、代碼生成)比較不同模型的優劣,同時揭示當前技術的瓶頸(如邏輯處理、穩定性)。

- **Reddit 連結**: [https://reddit.com/r/ClaudeAI/comments/1jpj7cs/claude_37_sonnet_is_still_the_best_llm_by_far_for/](https://reddit.com/r/ClaudeAI/comments/1jpj7cs/claude_37_sonnet_is_still_the_best_llm_by_far_for/)
- **外部連結**: [https://medium.com/p/f180b9c12bc1](https://medium.com/p/f180b9c12bc1)
- **發布時間**: 2025-04-02 15:49:40

### 內容

You're evaluating LLM capabilities through their ability to generate SEO-optimized, visually appealing websites. However, "Frontend development" is way more than just HTML and CSS.

What you're testing is essentially the surface-level appearance of websites - something that most modern LLMs can indeed generate in a single shot. This is only scratching the surface of what frontend development actually involves.

True frontend development includes:

  • JavaScript functionality and interactivity

  • State management

  • API integration

  • Performance optimization

  • Accessibility compliance

  • Build systems and tooling

  • Testing and debugging

While LLMs can generate decent-looking static templates, they struggle with the complex logic, architecture decisions, and technical problem-solving that professional frontend developers handle daily.

If you're genuinely interested in evaluating LLM capabilities in this domain, consider testing for more comprehensive aspec``` of frontend development beyond just visual appearance and SEO markup.

Claude 3.7 has become ass at coding. It cant even give me a method with identation like : " def..." I said to use propert spacing 10 times and it says "here is the method with proper spacing"

I've been the biggest Claude Sonnet fanboi for ages, but Gemini 2.5 Pro is better. It's so much better that it has solved problems not even Claude could solve and so much cheaper.

Gemini 2.5 is powerful, but it's so chaotic.


### 討論

**評論 1**:

You're evaluating LLM capabilities through their ability to generate SEO-optimized, visually appealing websites. However, "Frontend development" is way more than just HTML and CSS.

What you're testing is essentially the surface-level appearance of websites - something that most modern LLMs can indeed generate in a single shot. This is only scratching the surface of what frontend development actually involves.

True frontend development includes:

  • JavaScript functionality and interactivity
  • State management
  • API integration
  • Performance optimization
  • Accessibility compliance
  • Build systems and tooling
  • Testing and debugging

While LLMs can generate decent-looking static templates, they struggle with the complex logic, architecture decisions, and technical problem-solving that professional frontend developers handle daily.

If you're genuinely interested in evaluating LLM capabilities in this domain, consider testing for more comprehensive aspec``` of frontend development beyond just visual appearance and SEO markup.


**評論 2**:

Claude 3.7 has become ass at coding. It cant even give me a method with identation like : " def..." I said to use propert spacing 10 times and it says "here is the method with proper spacing"


**評論 3**:

I've been the biggest Claude Sonnet fanboi for ages, but Gemini 2.5 Pro is better. It's so much better that it has solved problems not even Claude could solve and so much cheaper.


**評論 4**:

Gemini 2.5 is powerful, but it's so chaotic.


---

## 11. ```
Pour one out for my Claude subscription... It's not you, it's Gemini (and my PhD).
``` {#11-```
pour-one-out-for-my-claude-subscription-it-}

這篇文章的核心討論主題是:
**作者從長期使用Claude AI轉向Gemini AI的過程,反映對兩款AI工具性能變化的體驗與情感糾結**。

具體要點包括:
1. **Claude的優勢**:作者(一名博士生)高度讚賞Claude對複雜研究任務的理解力與人性化互動,形容它像「永不偷吃零食的聰明實驗室夥伴」。
2. **Gemini的初期定位**:最初僅用於簡單查詢任務,作者仍優先保留Claude訂閱以處理高難度需求。
3. **轉折點**:最新版Gemini性能大幅提升,打破原有局限,促使作者取消Claude訂閱。
4. **情感矛盾**:決策伴隨不捨,作者表達對Claude的懷念(如「真正的朋友」),並隱晦批評其當前版本可能受限(「被關在籠子裡」的隱喻)。

本質上,這是一篇結合工具效能評估與個人情感的「AI服務遷移」經驗談。

- **Reddit 連結**: [https://reddit.com/r/ClaudeAI/comments/1jppakr/pour_one_out_for_my_claude_subscription_its_not/](https://reddit.com/r/ClaudeAI/comments/1jppakr/pour_one_out_for_my_claude_subscription_its_not/)
- **外部連結**: [https://www.reddit.com/r/ClaudeAI/comments/1jppakr/pour_one_out_for_my_claude_subscription_its_not/](https://www.reddit.com/r/ClaudeAI/comments/1jppakr/pour_one_out_for_my_claude_subscription_its_not/)
- **發布時間**: 2025-04-02 22:01:08

### 內容

Okay, fellow AI wranglers, confession time. For the longest time, Claude was the one. As a PhD student navigating the treacherous waters of research, Claude wasn't just smart; it got me. Frustrated ramblings? Check. Complex concep```? Handled. It was like having a super-intelligent, patient lab partner who never stole my snacks.

I even had a Gemini sub on the side, but 's be real Gemini got the simple stuff, the lookup tasks. My precious Claude credi were reserved for the real brain-busters, the momen where only Claude's uncanny understanding would do.

But then... the latest Gemini stepped up i``` game. Big time. Suddenly, the performance is stellar, and the limitations feel... well, gone from my workflow.

So, with a heavy heart (and a slightly lighter wallet), I'm cancelling my Claude subscription. I know my 22/month won't exactly bankrupt Anthropic, it's a drop in their massive ocean. But man, I'll miss that connection.

Farewell for now, Claude. You were a true friend and a helping hand during some tough research momen```. Here's hoping I can someday come back to a Claude that's not in a cage.


### 討論

**評論 1**:

Yup, you and me both. I'm paying $20/month and can't have a conversation longer than the one I have with the postman? Doesn't feel dependable when I could get shut out at in time in the middle of a working collaboration.


**評論 2**:

What is your workflow? Also am in the midst of doctoral studies but I am mostly using Claude to help me rephrase/format my nonsensical ramblings in professional emails and memos. With a very specific and refined prompt, I am able to get it to sound fairly natural and similar to my writing style. Is Gemini better than Claude for this now? I use NotebookLM mostly to keep track of literature and ask targeted questionsthis is still the best in that context.


**評論 3**:

Quick question, when you cancel your subscription does it kick you out immediately or do you at least get to run out your remaining days since your last renewal?


**評論 4**:

I also cancelled my Claude subscription


**評論 5**:

I continue to use Gemini 2.5 (subscription) and continue to be astounded at just how damn shitty it is. I mean, it can't even build a simple interactive dashboard worth a shit. I just aggregated a whole bunch of data concerning marijuana legalization, and asked Claude to turn it into an interactive dashboard which it did very well.

Gemini gave me... Well... Something. Not quite sure what the f it was. A couple of bizarre geometric shapes with seemingly random numbers written all over it. I have now done probably 15 maybe 20 interactive dashboards over the past couple of days with data I've collected, and literally in every instance Claude produced something far superior.

The latest one I did was simply aggregating the data for the 1993 NBA championship series to exemplify the amazing performance Michael Jordan had. Claude gave me four to five pages to click on, at least six or seven interactive char```, active data.

Gemini gave me one page, one chart, inaccurate data, and the data wouldn't even load.

Then I try to turn a bunch of data into a CSV file and configure it to fit into HubSpot. Claude knocks it out without much of a problem. Gemini 2.5 star spinning in circles and I think smoke started coming out of i electronic years. Kept giving me excuses why It can't do it.

Then I ask some reasonable questions of Claude, And I get pretty decent well thought out well structured answers. Gemini 2.5 proceeds to give me a Wikipedia page for everything. Every damn thing I ask it gives me three or four potential answers with each answer having multiple bullet poin```, it's like it's trying to write a damn textbook.

Then there's the search feature or I guess you can call that, if you like getting inaccurate information on virtually everything. I mean I must be in some parallel universe. Or maybe I'm just unlucky or something. But for virtually every use case I've tried it for it's been garbage.

Claude certainly isn't perfect, but out of all the llms I have subscriptions to (which is all of them) I always end up coming back to Claude for virtually 85% of the junk I do.


---

## 12. ```
Claude Code was prohibitively expensive for me
``` {#12-```
claude-code-was-prohibitively-expensive-for}

這篇文章的核心討論主題是:**作者對Claude Code(高階AI編碼工具)的效能與成本的權衡評估**,主要圍繞以下幾個重點:

1. **效能優勢**
- 解決了其他模型(如Sonnet 3.7)難以處理的問題
- 首次嘗試即成功,表現精準("surgical"修改)且具創造性
- 提供類似人類的創新解決方案

2. **成本顧慮**
- 每小時$21.75的費用對自由工作者而言過高(直接影響時薪收入)
- 對比固定月費工具(如Cursor的$40/月)顯得不划算

3. **未來期望**
- 希望降價以作為備用方案
- 期待客戶提供API金鑰(降低自身成本,同時為客戶節省開發時間)
- 對CLI工具操作體驗的肯定與改進空間

總結:作者在肯定Claude Code技術能力的同時,糾結於其現行定價模式與個人經濟效益的衝突,並提出潛在的雙贏解決方案。

- **Reddit 連結**: [https://reddit.com/r/ClaudeAI/comments/1jpihs2/claude_code_was_prohibitively_expensive_for_me/](https://reddit.com/r/ClaudeAI/comments/1jpihs2/claude_code_was_prohibitively_expensive_for_me/)
- **外部連結**: [https://www.reddit.com/r/ClaudeAI/comments/1jpihs2/claude_code_was_prohibitively_expensive_for_me/](https://www.reddit.com/r/ClaudeAI/comments/1jpihs2/claude_code_was_prohibitively_expensive_for_me/)
- **發布時間**: 2025-04-02 14:57:24

### 內容

At the rate I was using it, it would cost $21.75 per hour. It did an impressive job and solved a problem that other models (including Sonnet 3.7) were struggling with, and did so with i``` first attempt.

I haven't tried it more because of the expense. As a freelancing AI Engineer, that would be coming straight out of my hourly rate. Unlike Cursor, which I pay a fixed $40/month for.

I hope it will come down in cost, as it's nice to have a backup strategy. Some clien may provide me with an Anthropic key (the modern equivalent of providing a desk and chair), and then everyone wins because it would reduce the time it takes me to build AI produc, so a saving for them.

Looking forward to using it more. There's something reassuring about using CLI tools, though you have to jump into your IDE to review what was changed.

Claude Code was surgical and only made the minimum amount of changes. I``` solution was quite creative; it had taken a step back from the task to think about it in a new and novel way; a bit human-like in that regard, and with a good result.


### 討論

**評論 1**:

Have you tried Cline and Roo Code?


**評論 2**:

utilise le MCP jetbrains tu feras "a peu prs" la meme chose qu'avec Claude Code


**評論 3**:

We created an MCP Server that is as powerful as Claude Code but can be used completely for free through Claude Desktop:

https://github.com/oraios/serena


**評論 4**:

IMO the agen are still a ways off from being super useful given the cos to use them vs single api calls / cha and copy pasting giving you some majority % of the benefi and still helping your velocity.

But Claude code is among the worst offenders, as are cline / roo to some extent, though theyre a little better.

They do almost zero optimizations for context management and throw the full project into every query (cline and roo throw the full touched files iirc) and it balloons cos for it. It also can hurt performance as the projec grow because coherence on medium projec (\>4k lines? 5-15 tokens for a line is reasonable depending on the language) pu you at a 32k window which is almost unusable for most models.

In theory cos``` will come down a lot as we optimize for them and get better RAG retrieval, but for first party solutions - Claude code, oai / googles eventual one, etc they have perverse incentives to do so.

Aiders system is pretty cool - using what they call a repository map that traverses the AST instead of raw vector embeddings. Continues is neat because of how customizable it is - rag is default but you have providers for @open / @codebase etc to override it when needed. I havent used their agent - but the plugin is reasonably useful.


**評論 5**:

How does claude code differ from just using the chat sessions?


---

## 13. ```
Claude created this game prototype for an idea I had in a single attempt. Now I've turned it into a full fledged game
``` {#13-```
claude-created-this-game-prototype-for-an-i}

從提供的對話片段來看,核心討論主題似乎圍繞著「遊戲開發與發布」的過程或成果。以下是具體分析:

1. **遊戲創作與合作**
「he helped me make the same game」暗示有人協助完成了一款遊戲的開發,可能涉及合作或技術支援。

2. **遊戲發布的興奮或調侃**
「Vibe it onto the App Store!」帶有輕鬆或幽默的語氣,可能是對將遊戲上架應用商店的期待或玩笑。

3. **對現有遊戲的評價**
「This is already a game. Still cool」可能指該遊戲與現有作品相似,但仍被認為有趣,觸及原創性或重複性的討論。

4. **簡短回應的開放性**
「And?」可能表達對進一步討論的期待,例如後續計劃、反饋,或對話題的冷淡回應。

**總結**:對話的核心主題是「遊戲開發的協作、發布及相關反饋」,內容涉及從製作到上架的過程,並夾雜對遊戲相似性或價值的簡短評論。語氣偏向非正式,可能出自開發者間的閒聊或社群互動。

- **Reddit 連結**: [https://reddit.com/r/ClaudeAI/comments/1jpbntl/claude_created_this_game_prototype_for_an_idea_i/](https://reddit.com/r/ClaudeAI/comments/1jpbntl/claude_created_this_game_prototype_for_an_idea_i/)
- **外部連結**: [https://v.redd.it/i01286bghbse1](https://v.redd.it/i01286bghbse1)
- **發布時間**: 2025-04-02 08:38:10

### 內容

what a hussy, he helped me make the same game

Vibe it onto the App Store!

Very cool. This is already a game. Still cool

And?


### 討論

**評論 1**:

what a hussy, he helped me make the same game


**評論 2**:

Vibe it onto the App Store!


**評論 3**:

Very cool. This is already a game. Still cool


**評論 4**:

And?


---

## 14. ```
Claude's context has been nerfed?
``` {#14-```
claude-s-context-has-been-nerfed-
```}

这篇文章的核心討論主題是:**對Anthropic的Sonnet模型在長對話中「遺忘」程式碼片段現象的質疑與推測**。

作者遇到一個問題:在對話長度遠低於官方宣稱的200k token限制(約40k tokens)時,模型突然無法回憶起幾條訊息前的程式碼內容,即使反覆提示也無法正確重現或調整。這引發作者懷疑Anthropic可能未真實提供完整的上下文窗口,而是透過某種「摘要壓縮」機制來節省成本,具體推測如下:

1. **「摘要替代」假說**:
- 實際運作中,系統可能將過往訊息替換為高品質摘要(而非保留原始內容),導致模型雖名義上有200k上下文,但有效資訊量大幅縮水。
- 作者提出一套可能的技術流程(如哈希比對、動態摘要替換),推測此設計會犧牲回應品質以降低運算成本。

2. **排除其他可能性**:
- 若因模型量化(quantization)導致效能下降,應表現為整體回應品質漸進退化,而非特定內容的突然遺忘。
- 當前現象更符合「選擇性摘要」造成的資訊缺失。

3. **商業動機質疑**:
- 作者暗示此做法可能是變相降低服務成本的手段(「節省萬億dollaroos」),同時維持「200k上下文」的宣傳噱頭。

總結:文章核心在於批判Anthropic可能透過隱蔽的摘要機制壓縮實際上下文,導致模型在長對話中出現非預期的遺忘問題,並質疑其透明度與真實效能承諾。

- **Reddit 連結**: [https://reddit.com/r/ClaudeAI/comments/1jp4u8u/claudes_context_has_been_nerfed/](https://reddit.com/r/ClaudeAI/comments/1jp4u8u/claudes_context_has_been_nerfed/)
- **外部連結**: [https://www.reddit.com/r/ClaudeAI/comments/1jp4u8u/claudes_context_has_been_nerfed/](https://www.reddit.com/r/ClaudeAI/comments/1jp4u8u/claudes_context_has_been_nerfed/)
- **發布時間**: 2025-04-02 03:44:51

### 內容

Like every day, I was doing some coding, with 3.7 and things were going swellingly and then suddenly, a chunk of code, that was roughly 50 locs long 2 message prior, was (almost) completely gone from Sonnet's context. The message from Sonnet was there, it referred to the like before, but despite ~10 edi to prompt it in different ways, it just couldn't re-produce the nor make correct adjustmen to the . Aside from knowing the name and parameters, it seemed to be clueless.

The conversation is well below the 200k token limit, at around 40k tokens, which makes me wonder how it's even possible? If the model got quantized to shit, it wouldn't just completely lose context. It would give worse responses, but 2 messages back would be yielding better recollection than the initial message (which is not the case). Or alternatively the quality of responses would degrade to a repeating mess, but the "quality" felt exactly the same as prior to it. It just "forgot" the details.

So I'm wondering if what is happening is that they're using a sort of alternative prompt caching method (at least) for the chat client where the prior messages are collapsed into high quality summaries of previous assistant and user messages? Meaning, they're basically selling 200k context, but in reality it's 15k of summaries and you hit your limit @ 20k which in Anthropic math would be 200k* (*simulated tokens which are definitely worth exactly as much as real ones)?

Obviously this is just a tummy feel, but the above thing did happen and the only way for it to happen (as far as I can imagine) is either due to novel model degradation or the above very believable scam.

I reckon it would work something like this,

  1. Message comes in

  2. Generate a very high quality summary from the message (I'm sure they would've tested this to death)

  3. Store the summary with the hash of the message

  4. Generate completion to the original message

  5. Generate summary from completion

  6. Store the summary of the completion with the hash of the message

  7. New completion request arrives with the full message array

  8. Check the hashes of the messages and replace them with the summarized versions

  9. Return a more shitty completion.

  10. Save a trillion dollaroos on long conversations since every completion on long conversations is in reality now 90% shorter while selling a product that's been nerfed to death.

I doubt it would start from the first message, but there is some point where it becomes more economical to run 1 extra summarization on every message to get to perform completions on shorter conversations.


### 討論

**評論 1**:

Yeah... Claude's been doing this to me; Deepseek V3 will accept the promp``` and extra bloat context I shove sometimes; but nothing is beating Gemini's 1 Million Context window rn...


**評論 2**:

This happens to me daily, most often with MCP tool use, at some point where there is too much content it just breaks and the whole message disappears. I shared this as a bug via their feedback page but nothing has changed since.


**評論 3**:

I'm experiencing similar context limitations, but in a different way! Before Claude 3.7 Sonnet was released, I (as a free user) could easily ask for title recommendations for my 2000-2500 word English articles. Claude would read through the entire text and provide thoughtful title suggestions based on the full content.

But now? It seems completely nerfed. I can only get Claude to properly handle articles up to about 1500 words max. Anything longer and the system won't even me send the message at all. I get an error saying: "Your message will exceed the length limit for this chat. Try attaching fewer or smaller files or starting a new conversation."


**評論 4**:

Yes -- there are 'holes' in the context window. I've seen this with Claude and with some GPT-4 varian``` as well. I've also observed it while playing around with smaller local models on my own GPU, where the max context is smaller and it kicks in much sooner for some models. Sometimes by posting another message or two you can shift the 'hole' somewhere else, sometimes. Usually you'll need to start over with a new session

I do think this is less 'something they did on purpose' and more 'attention heads have weird blind-spo``` with certain content and nobody can quite untangle what and why'

It is very frustrating when it happens


---

## 15. ```
Anthropic, come get your boy. Cursor + Claude 3.7 MAX
``` {#15-```
anthropic-come-get-your-boy-cursor-claude-3}

這篇文章的核心討論主題是**提交效能證明(proof of performance)時的必要要求與規範**,具體包括:

1. **提交內容的強制性要求**
- 需附上輸出結果的截圖
- 提供完整的提示詞(prompt)序列(若相關)
- 註明使用環境(免費/付費網頁版或API)

2. **違反規定的後果**
- 文章可能被刪除或重新標記分類(flair)
- 鼓勵用戶主動檢舉不符合規定的貼文

3. **適用情境**
主要針對需要驗證AI模型(如ClaudeAI)表現的技術討論場景,強調透明性與可重現性。

(註:文中部分單詞如「screensho```」「promp```」應為「screenshot」「prompt」的輸入錯誤)

- **Reddit 連結**: [https://reddit.com/r/ClaudeAI/comments/1jpdw34/anthropic_come_get_your_boy_cursor_claude_37_max/](https://reddit.com/r/ClaudeAI/comments/1jpdw34/anthropic_come_get_your_boy_cursor_claude_37_max/)
- **外部連結**: [https://i.redd.it/918gsfnw0cse1.jpeg](https://i.redd.it/918gsfnw0cse1.jpeg)
- **發布時間**: 2025-04-02 10:24:21

### 內容

When submitting proof of performance, you must include all of the following:

  1. Screensho``` of the output you want to report

  2. The full sequence of promp``` you used that generated the output, if relevant

  3. Whether you were using the FREE web interface, PAID web interface, or the API if relevant

If you fail to do this, your post will either be removed or reassigned appropriate flair.

Please report this post to the moderators if does not include all of the above.

I am a bot, and this action was performed automatically. Please contact the moderators of this subreddit if you have any questions or concerns.


### 討論

**評論 1**:

When submitting proof of performance, you must include all of the following:

  1. Screensho``` of the output you want to report
  2. The full sequence of promp``` you used that generated the output, if relevant
  3. Whether you were using the FREE web interface, PAID web interface, or the API if relevant

If you fail to do this, your post will either be removed or reassigned appropriate flair.

Please report this post to the moderators if does not include all of the above.

I am a bot, and this action was performed automatically. Please contact the moderators of this subreddit if you have any questions or concerns.


---

## 16. ```
Now we talking INTELLIGENCE EXPLOSION | Claude 3.5 cracked of benchmark!
``` {#16-```
now-we-talking-intelligence-explosion-|-cla}

這段討論的核心主題是 **不同AI模型(Claude 3.5 Sonnet vs. OpenAI o-series)在兩種代理框架(BasicAgent vs. IterativeAgent)下的性能差異與潛在優化方向**,具體要點如下:

1. **模型表現對比**
- Claude 3.5 Sonnet 在 **BasicAgent**(允許自主決定任務結束)設定下顯著優於其他模型(20% vs. OpenAI o1-high的13%)。
- 但在 **IterativeAgent**(強制分階段執行,禁止提前結束)設定下,OpenAI o1-high 表現最佳(25%),而Claude 3.5 降至16%。

2. **關鍵發現**
- Claude 3.5 在自主決策時表現更強,而其他模型(如OpenAI)容易過早聲稱完成任務或放棄。
- 現有的IterativeAgent提示設計可能更適合OpenAI模型,暗示Claude 3.5的潛力未被充分釋放。

3. **優化方向**
- 調整代理框架(例如修改BasicAgent以禁止提前結束)可能讓Claude 3.5超越OpenAI的表現。
- 當前基準測試的結果高度依賴提示設計和代理架構,存在「低垂果實」(易實現的改進空間)。

4. **延伸討論**
- 提及GPT-3.7(假設性版本)可能因更強的自主性(agentic)和思考能力而表現不同,與GPT-3.5的迭代限制形成對比。

**總結**:討論聚焦於不同AI模型在任務執行框架中的適應性差異,並強調提示工程(prompt tuning)和代理設計對基準測試結果的關鍵影響,同時指出Claude 3.5的優化潛力。

- **Reddit 連結**: [https://reddit.com/r/ClaudeAI/comments/1jpubua/now_we_talking_intelligence_explosion_claude_35/](https://reddit.com/r/ClaudeAI/comments/1jpubua/now_we_talking_intelligence_explosion_claude_35/)
- **外部連結**: [https://i.redd.it/upvmyq9shgse1.jpeg](https://i.redd.it/upvmyq9shgse1.jpeg)
- **發布時間**: 2025-04-03 01:25:34

### 內容

Would be interesting to see it with 3.7 since it's a lot more agentic compared to 3.5 and also has thinking. Iteration actually seemed to hamper 3.5 in this case:

https://imgur.com/a/evhqUgu

Here's the paper link btw:

https://cdn.openai.com/papers/22265bac-3191-44e5-b057-7aaacd8e90cd/paperbench.pdf

What's really interesting, from the paper, is Claude 3.5 destroys everyone in the "BasicAgent" set up ~20% to o1-high's 13% or o3mini-high's 3%.

But using their "IterativeAgent" setup which "removes the ability of the models to end the task early and promp models to work in a piecemeal fashion" o1-high se a new record at ~25% but Claude 3.5 drops to 16%.

"We observe that all models apart from Claude 3.5 Sonnet frequently finished early, claiming that they had either finished the entire replication or had faced a problem they couldn't solve"

"We note with Claude 3.5 Sonnet outperforms o1 with Basic Agent but underperforms o1 with IterativeAgent. This suggest that the prompt tuning used for iterative agent is differently suited for OpenAI o-series models. We suspect that a modification to BasicAgent that also preven``` it from ending the task early could lead to Claude 3.5 Sonnet outperforming o1 with IterativeAgent"

Sounds like there are still easy gains in this benchmark just from tweaking initial prompt setup and agent framework.


### 討論

**評論 1**:

Would be interesting to see it with 3.7 since it's a lot more agentic compared to 3.5 and also has thinking. Iteration actually seemed to hamper 3.5 in this case:
https://imgur.com/a/evhqUgu

Here's the paper link btw:
https://cdn.openai.com/papers/22265bac-3191-44e5-b057-7aaacd8e90cd/paperbench.pdf


**評論 2**:

What's really interesting, from the paper, is Claude 3.5 destroys everyone in the "BasicAgent" set up ~20% to o1-high's 13% or o3mini-high's 3%.

But using their "IterativeAgent" setup which "removes the ability of the models to end the task early and promp models to work in a piecemeal fashion" o1-high se a new record at ~25% but Claude 3.5 drops to 16%.

"We observe that all models apart from Claude 3.5 Sonnet frequently finished early, claiming that they had either finished the entire replication or had faced a problem they couldn't solve"

"We note with Claude 3.5 Sonnet outperforms o1 with Basic Agent but underperforms o1 with IterativeAgent. This suggest that the prompt tuning used for iterative agent is differently suited for OpenAI o-series models. We suspect that a modification to BasicAgent that also preven``` it from ending the task early could lead to Claude 3.5 Sonnet outperforming o1 with IterativeAgent"

Sounds like there are still easy gains in this benchmark just from tweaking initial prompt setup and agent framework.


---

## 17. ```
What Happens When You Tell an LLM It Has an iPhone Next to It?
``` {#17-```
what-happens-when-you-tell-an-llm-it-has-an}

上述討論的核心主題可總結為以下三點:

1. **實驗設計的嚴謹性**
首段強調控制組(control groups)的重要性,指出需比較不同條件(無提示、財務分析師角色、智能手機提示、混合提示)才能得出可靠結論,並延伸探討「不同物件」是否會影響人類或AI的反應差異。

2. **提示工程(Prompt Engineering)的影響機制**
後兩段聚焦於「用戶訊息(User Message)」的附加內容如何改變模型輸出。討論提出假設:
- 在用戶訊息中追加內容(如角色設定或上下文)會顯著影響結果,因其擴大了模型的「搜索窗口(search window)」,提供更多關聯線索。
- 反之,若僅在系統提示(system prompt)中添加內容,而非用戶訊息,則效果可能有限。

3. **對AI行為的推測與驗證需求**
整體討論體現對AI模型回應機制的興趣,包括:
- 不同提示策略(如物件、角色設定)的比較。
- 需進一步實驗驗證假設(例如移除用戶訊息追加內容的對照測試)。

**關鍵詞彙**:控制組、提示工程、用戶訊息與系統提示的差異、上下文窗口、AI行為可解釋性。

- **Reddit 連結**: [https://reddit.com/r/ClaudeAI/comments/1jpr724/what_happens_when_you_tell_an_llm_it_has_an/](https://reddit.com/r/ClaudeAI/comments/1jpr724/what_happens_when_you_tell_an_llm_it_has_an/)
- **外部連結**: [https://medium.com/p/01a82c880a56](https://medium.com/p/01a82c880a56)
- **發布時間**: 2025-04-02 23:19:29

### 內容

I'm not a scientist or anything, but you'd need more control groups. One with no additional prompting, one with the financial analyst persona, one with the smartphone and one with both.

Without it, it's kind of hard to draw a proper conclusion.

I also wonder if it being a different object would make a difference in humans or AI in this case.

This is cool

It genuinely seems like an interesting thought experiment. My guess is that appending the User Message is where the actual result transformation is taking place. Id be interested to see the changes in resul``` without appending the user message.

By adding tokens to the User Message the model has more context to explore what the User is requesting, correlating to a large search window. The larger the window, the more you can see, the more accurate the picture can be. That makes sense to me given the resul```

My guess is that if you stop appending the user message, and simply add a few lines to a gigantic system prompt it doesnt change as much, if at all.


### 討論

**評論 1**:

I'm not a scientist or anything, but you'd need more control groups. One with no additional prompting, one with the financial analyst persona, one with the smartphone and one with both.
Without it, it's kind of hard to draw a proper conclusion.

I also wonder if it being a different object would make a difference in humans or AI in this case.


**評論 2**:

This is cool


**評論 3**:

It genuinely seems like an interesting thought experiment. My guess is that appending the User Message is where the actual result transformation is taking place. Id be interested to see the changes in resul``` without appending the user message.

By adding tokens to the User Message the model has more context to explore what the User is requesting, correlating to a large search window. The larger the window, the more you can see, the more accurate the picture can be. That makes sense to me given the resul```

My guess is that if you stop appending the user message, and simply add a few lines to a gigantic system prompt it doesnt change as much, if at all.


---

## 18. ```
The Impact of Generative AI on Critical Thinking - Research Paper
``` {#18-```
the-impact-of-generative-ai-on-critical-thi}

該文章的核心討論主題為 **「生成式人工智慧(GenAI)對批判性思維的影響」**,主要探討以下幾個關鍵面向:

1. **技術依賴性與自主性**
- GenAI的普及可能導致知識工作者過度依賴技術,但批判性思維能力較高者更傾向獨立評估AI輸出或自主完成任務。

2. **批判性思維的轉變**
- GenAI改變傳統認知流程:從「資訊蒐集」轉向「資訊驗證」,從「直接解決問題」轉向「整合AI回應」,同時增加驗證AI結果所需的認知努力。

3. **信心與批判性思維的關聯**
- 使用者對自身能力的信心及對AI的信任度,會影響其批判性思維的投入程度;過度依賴AI可能削弱批判性思考。

4. **角色轉變:從執行到監督**
- AI自動化例行任務(如資料蒐集、初稿生成)後,工作者需轉型為「任務管理者」,專注於監督、驗證和優化AI產出以符合標準。

5. **促進與阻礙批判性思維的因素**
- 動機(如提升工作品質、避免錯誤)推動批判性思考,而缺乏意識、動機或技能則成為主要障礙,需透過干預措施改善。

**總結**:文章聚焦於GenAI如何重塑人類的認知行為模式,強調在AI輔助工作中維持批判性思維的平衡,並提出技術設計與工作流程優化的潛在方向。

- **Reddit 連結**: [https://reddit.com/r/ClaudeAI/comments/1jpd1pm/the_impact_of_generative_ai_on_critical_thinking/](https://reddit.com/r/ClaudeAI/comments/1jpd1pm/the_impact_of_generative_ai_on_critical_thinking/)
- **外部連結**: [https://www.reddit.com/r/ClaudeAI/comments/1jpd1pm/the_impact_of_generative_ai_on_critical_thinking/](https://www.reddit.com/r/ClaudeAI/comments/1jpd1pm/the_impact_of_generative_ai_on_critical_thinking/)
- **發布時間**: 2025-04-02 09:46:34

### 內容

The Impact of Generative AI on Critical Thinking

Research Insigh```:

  • Technology Over-Reliance: The research indicates a potential increase in technology over-reliance with accessible GenAI tools among knowledge workers. The study shows that higher critical thinking levels were associated with greater confidence in performing tasks independently of AI or in evaluating AI responses, suggesting a correlation between self-reliance and critical engagement.

  • Critical Thinking Transition: A significant transition is observed in critical thinking with GenAI use, as the focus shif from mere information gathering to rigorous information verification, and from traditional problem-solving to integrating AI responses. Empirical evidence sugges that while GenAI tools reduce the perceived effort for cognitive activities like knowledge retrieval, they simultaneously necessitate increased effor in verifying AI-generated resul to ensure quality and accuracy.

  • Confidence and Critical Thinking: Users task-specific self-confidence and their confidence in GenAIs capabilities predict the occurrence and effort of critical thinking. Higher reliance on GenAI tends to correlate with diminished critical thinking, whereas those with enhanced self-confidence exhibit a greater propensity for critical engagement, highlighting the intricate balance between trust in AI and autonomous reasoning.

  • Task Stewardship: The study poin to a transition from task execution to task stewardship, implying a role shift towards overseeing AI to ensure the production of high-quality work. While AI proves instrumental in automating routine tasks such as information gathering and initial content creation, it consequently demands new effor in verifying and fine-tuning AI outpu``` to align with specific criteria and standards.

  • Motivators and Barriers for Critical Thinking: Motivators for critical thinking support in AI-assisted work are driven by the need to enhance work quality and avert potential negative consequences. Barriers such as a lack of awareness, motivation, and skills necessary for improving AI outpu``` pose as significant inhibitors, suggesting potential areas for intervention in technology-autonomous workflows.


### 討論

**評論 1**:

In before the trusty "so calculators make everyone dumb too then!" argument


**評論 2**:

bro used Claude to write a summary on a paper talking about the erosion of critical thinking skills using GenAI. impressive meta


---

## 19. ```
Claude 3.7 Sonnet extended thinking is a waste of time (for deep coding) since Monday
``` {#19-```
claude-3-7-sonnet-extended-thinking-is-a-wa}

這篇文章的核心討論主題是:**用戶對於系統重複出現相同錯誤(在此例中為明顯的重複內容)的困惑與不滿**,並詢問其他用戶是否也遇到類似問題。

重點包括:
1. **問題描述**:用戶指出系統在多次反饋後仍重複出現相同的錯誤(如內容重複)。
2. **時間對比**:用戶提到此類錯誤在上週並未發生,暗示近期可能出現異常。
3. **尋求共鳴**:詢問其他用戶是否遇到類似情況,以確認是否為普遍現象或系統性問題。

整體而言,這是一個關於**系統錯誤重現的質疑與社群反饋的請求**。

- **Reddit 連結**: [https://reddit.com/r/ClaudeAI/comments/1jpq9nn/claude_37_sonnet_extended_thinking_is_a_waste_of/](https://reddit.com/r/ClaudeAI/comments/1jpq9nn/claude_37_sonnet_extended_thinking_is_a_waste_of/)
- **外部連結**: [https://www.reddit.com/r/ClaudeAI/comments/1jpq9nn/claude_37_sonnet_extended_thinking_is_a_waste_of/](https://www.reddit.com/r/ClaudeAI/comments/1jpq9nn/claude_37_sonnet_extended_thinking_is_a_waste_of/)
- **發布時間**: 2025-04-02 22:41:39

### 內容

https://preview.redd.it/8lwji1e6ofse1.png?width=491&format=png&auto=webp&s=476ed5b5d7aec8c814613795cbc50dcdcdafc0b7

I have received this obviously duplicated after 2 iterations of pointing this problem out.

Mistakes like this didn't happen last week, which I find strange.

Anyone else seeing similar problems?


### 討論

**評論 1**:

Since a while, I switched to code reviews with o3 mini high to slap a bit Sonnet 3.7.
And that helps steer the dumb thinking.
Sonnet remain a very solid working horse.


**評論 2**:

It actually happened on Saturday in my experience they ruined it somehow by tuning: it still produces code that compiles, but doesn't solve your problem anymore. I was so pissed yesterday with it, it became worse than Gemini 2.0 flash


---

## 20. ```
Dropped a new tutorial for Agentic pattern + AI SDK
``` {#20-```
dropped-a-new-tutorial-for-agentic-pattern-}

這段文字的核心討論主題是:

1. **介紹新影片內容**:創作者發布了一支關於「Agentic patterns」(能動性模式)的新影片,並提到內容參考了Anthropic的研究論文。
2. **尋求觀眾反饋**:創作者強調這是他的第三支影片,希望觀眾提供意見以改進未來作品,包括對影片內容、表現方式的看法。
3. **互動與成長**:表達了對觀眾參與的期待(如留言反饋),並展現持續進步的意願。

簡單總結:**「分享Agentic patterns相關影片,並誠摯請求觀眾反饋以優化創作」**。

- **Reddit 連結**: [https://reddit.com/r/ClaudeAI/comments/1jp6sai/dropped_a_new_tutorial_for_agentic_pattern_ai_sdk/](https://reddit.com/r/ClaudeAI/comments/1jp6sai/dropped_a_new_tutorial_for_agentic_pattern_ai_sdk/)
- **外部連結**: [https://www.reddit.com/r/ClaudeAI/comments/1jp6sai/dropped_a_new_tutorial_for_agentic_pattern_ai_sdk/](https://www.reddit.com/r/ClaudeAI/comments/1jp6sai/dropped_a_new_tutorial_for_agentic_pattern_ai_sdk/)
- **發布時間**: 2025-04-02 05:02:10

### 內容

Hey guys, I just dropped a new video covering Agentic patterns. I'll be covering all the Agentic patterns that are commonly used using the Anthropic paper.

Would really love your though``` on what you think about this video and whether I can improve. This is only my third video, and I will get better but could really use some feedback.

https://youtu.be/KE8jb6adxUQ


### 討論

無討論內容

---

## 21. ```
This conversation reached i``` maximum length
``` {#21-```
this-conversation-reached-i```-maximum-leng}

這篇文章的核心討論主題是:**用戶對Pro方案服務品質突然下降的抱怨**,特別是近期出現的**使用限制過早觸發的問題**(例如僅閱讀幾章內容後就遇到限制),與過去幾個月穩定體驗的對比。

具體焦點包括:
1. **異常現象**:長期正常使用後,近期無預警出現限制觸發過早的問題。
2. **付費用戶的不滿**:作為付費(Pro方案)用戶,對服務可靠性的質疑。
3. **尋求共鳴**:詢問其他用戶是否遇到類似情況,暗示問題可能具有普遍性。

- **Reddit 連結**: [https://reddit.com/r/ClaudeAI/comments/1jpnrn3/this_conversation_reached_its_maximum_length/](https://reddit.com/r/ClaudeAI/comments/1jpnrn3/this_conversation_reached_its_maximum_length/)
- **外部連結**: [https://www.reddit.com/r/ClaudeAI/comments/1jpnrn3/this_conversation_reached_its_maximum_length/](https://www.reddit.com/r/ClaudeAI/comments/1jpnrn3/this_conversation_reached_its_maximum_length/)
- **發布時間**: 2025-04-02 20:50:06

### 內容

Is anyone noticing that they are hitting this way too soon the last few days? I have been using Pro plan for a few months and never had this issue. Suddenly this happens only after a couple of chapters.


### 討論

**評論 1**:

When making a complaint, please

  1. make sure you have chosen the correct flair for the Claude environment that you are using: i.e Web interface (FREE), Web interface (PAID), or Claude API. This information helps others understand your particular situation.
  2. try to include as much information as possible (e.g. prompt and output) so that people can understand the source of your complaint.
  3. be aware that even with the same environment and inpu```, others might have very different outcomes due to Anthropic's testing regime.
  4. be sure to thumbs down unsatisfactory Claude output on Claude.ai. Anthropic representatives tell us they monitor this data regularly.

I am a bot, and this action was performed automatically. Please contact the moderators of this subreddit if you have any questions or concerns.


**評論 2**:

How are you handling memory storage? Project functions, Google docs/shee``` or MCP/other?


---

## 22. Anthropic is giving free API credi``` for university studen``` \{#22-anthropic-is-giving-free-api-credi```-for-unive}

這篇文章的核心討論主題是提供學生開發者申請與Anthropic公司聯繫的申請表連結。重點在於說明如何透過指定的網址(`https://www.anthropic.com/contact-sales/for-student-builders`)獲取相關資源或支援。

簡要總結:
**「提供學生開發者專用的Anthropic銷售聯繫申請表連結」**

- **Reddit 連結**: [https://reddit.com/r/ClaudeAI/comments/1jpwevh/anthropic_is_giving_free_api_credits_for/](https://reddit.com/r/ClaudeAI/comments/1jpwevh/anthropic_is_giving_free_api_credits_for/)
- **外部連結**: [https://www.reddit.com/r/ClaudeAI/comments/1jpwevh/anthropic_is_giving_free_api_credits_for/](https://www.reddit.com/r/ClaudeAI/comments/1jpwevh/anthropic_is_giving_free_api_credits_for/)
- **發布時間**: 2025-04-03 02:46:55

### 內容

Application form is here: https://www.anthropic.com/contact-sales/for-student-builders


### 討論

無討論內容

---

## 23. ```
I'm not having issues?
``` {#23-```
i-m-not-having-issues-
```}

這段文章的核心討論主題可以總結為以下幾點:

1. **使用AI工具(如Claude)進行編程時的效率與限制**:
- 作者指出,直接要求AI生成完整程式或使用過於龐大的程式碼可能導致問題,建議從頭開始或建立可模組化的系統(如遊戲開發中的「plug and play」系統)以提高效率。
- 強調使用者需自行檢查生成的程式碼,避免盲目依賴AI。

2. **Claude的潛在問題與風險**:
- Claude可能過度「硬編碼」數值或虛構函式以滿足需求,導致程式碼看似可用但實際存在缺陷。
- 生成內容的可靠性與使用者輸入的明確性密切相關,模糊或簡單的提示可能導致低品質輸出。

3. **對使用者行為的批評與建議**:
- 批評部分使用者試圖用AI處理過大或複雜的任務,或完全依賴實驗性工具完成整個工作流程。
- 指出免費版網頁介面的效率低下,尤其缺乏專案管理功能時,建議改用API或調整使用方式(如分段請求「Continue」時需警惕輸出品質下降)。

4. **對「速率限制」的觀點**:
- 認為當前限制暴露了使用者對工具的誤解或過度依賴,呼籲更務實地評估AI的適用場景。

**總結**:文章主要探討如何有效且謹慎地使用AI編程工具(如Claude),強調使用者需明確需求、分段驗證輸出,並避免不切實際的依賴,同時指出工具本身的設計限制與潛在風險。

- **Reddit 連結**: [https://reddit.com/r/ClaudeAI/comments/1jpksic/im_not_having_issues/](https://reddit.com/r/ClaudeAI/comments/1jpksic/im_not_having_issues/)
- **外部連結**: [https://www.reddit.com/r/ClaudeAI/comments/1jpksic/im_not_having_issues/](https://www.reddit.com/r/ClaudeAI/comments/1jpksic/im_not_having_issues/)
- **發布時間**: 2025-04-02 17:52:02

### 內容

I've seen a lot of these pos, and it does make me think. I've noticed the downtime, I will not debate that, I've encountered it myself. But I do think the limi some of you are wondering about do confuse me, what are you prompting it with?

My though are that maybe trying to vibe code entire programs or using already bloated code might be part of the issue, combined with vague or simple promp. My experience is to use this you need to either start from scratch, or have a 'system' (game dev) in place that can support plug and play systems to get effective work done. You're always going to need to proof check this stuff.

As a result of all of this I've had no issue, nor a reliance dependency when it's down.

I will say that it does concern me, given that Claude needs to be specified not to hard code values & the extent of what I'm seeing people request it do without checking for these things. Claude is very very clever at making code work that shouldn't by inserting the values, or inventing fake functions for 'later use'.

My experience is the more times you've gotta type 'Continue', the sketchier the end product becomes. I wouldn't even attempt serious work without projec``` if I'm using the Web interface.

Tldr; this rate limit is exposing potential oversigh in people using this for 'too large' tasks or relying on an experimental product already for entire workflow solutions, people missing the API is intended for coding & the web interface is an incredibly inefficient method if doing it especially without projec if using free plan


### 討論

**評論 1**:

Until this point I haven't had any issues with Claude hard coding shit in because I still review every diff carefully to ensure it's not doing random shit. Also I think people are still copy and pasting into the webUI which makes Claude work like 10x worse. I'm surprised I dealt with it for so long. tldr: skill issues


**評論 2**:

Just sharing some though about Claude from my experience. It works okay if you use the MCP file system, add some custom instructions, and keep your promp straightforward.

Ive found its really only reliable for one task at a time. When I try to do multiple things, it star``` getting confused. The longer we talk, the more it seems to lose the plot. Thats why I usually try to get in, get what I need, and get out.

The last couple days have been particularly rough performance-wise, so Ive taken a step back from using Claude. Been checking out some other systems and LLMs instead.

Two things Ive noticed:

First, there seems to be a lot of wasted space in Claudes responses - Id say about 75% is just extra fluff with feature bloat offerings, and a lot of blatant refusing to follow directions. Whats really frustrating is when it finally gives you the perfect answer, and then it disappears with an error message, and you can never get that exact solution again.

Second, I wonder if theyre allocating resources differently for different customers. Do the big corporate clien``` get the full version while the rest of us get something more stripped down?

As a business owner, I read through all the policies and updates, and I have to say, their lack of transparency has made me hesitant to rely on them. I had similar trust issues with ChatGPT after they changed their policies. It just makes it hard to commit when things keep shifting without clear communication.


**評論 3**:

Claude works best if you create libraries for specific features then use those libraries to troubleshoot your code in new cha. Once it ge too monumental it has issues like any other AI.


**評論 4**:

Yeah, the entitlement those "contributions" reek of tell me that the respective OPs are probably not able to read the room or assess their own position in it. They unfortunately never got told by anybody that their complete lack of any skill and understanding is an actual reason they should stop doing things in the IT world and instead go back to cleaning the street, like it used to be in the past. And I'm actually not trying to hate on the cleaning professions, if they were properly compensated, etc bla bla...


---

## 24. ```
Claude Degradation Regionally Related?
``` {#24-```
claude-degradation-regionally-related-
```}

這段討論的核心主題是:**用戶對Claude(AI服務)性能問題的區域性差異探討**。

重點包括:
1. **性能抱怨的兩大途徑**:用戶透過API和聊天介面使用Claude時遇到問題,但並非所有人均受影響。
2. **區域與基礎設施的影響**:發文者身處美國大城市(德州),擁有高速網路,幾乎未遇問題,推測其他用戶的困境可能與地區性網路或服務部署有關。
3. **問題排查建議**:呼籲釐清是否為區域性問題(如伺服器負載、網路延遲),並尋求集體反饋以協助改善。

整體聚焦於「性能體驗差異」與「潛在區域因素」之間的關聯性探討。

- **Reddit 連結**: [https://reddit.com/r/ClaudeAI/comments/1jpb8mn/claude_degradation_regionally_related/](https://reddit.com/r/ClaudeAI/comments/1jpb8mn/claude_degradation_regionally_related/)
- **外部連結**: [https://www.reddit.com/r/ClaudeAI/comments/1jpb8mn/claude_degradation_regionally_related/](https://www.reddit.com/r/ClaudeAI/comments/1jpb8mn/claude_degradation_regionally_related/)
- **發布時間**: 2025-04-02 08:17:30

### 內容

Ive seen a lot of complain``` regarding Claudes performance both via API and Chat interface, and I sincerely dont think its across the board. Honestly, I have had zero issues (okay, maybe I have to resubmit an API call a second time every now and then, but its rare), but Im also in one of the largest cities in the USA. Ive got gigabit internet through fiber at work and home.

Im in TX, USA with zero issues when using the API.

I havent tested the chat interface, but Ive been reading pain poin``` from both avenues

Anyway, if we can locate the regional issue (if one exis), maybe we can make some noise in the correct way? I dunno. Just trying to help troubleshoot the shitorm Claude has become for many of you.


### 討論

**評論 1**:

Time of day matters a lot on the east coast so I believe it. The difference between 3am and 3pm performance is mind-blowing


**評論 2**:

honestly if they don't post screensho``` as a proof that Claude has degraded, you can just ignore those pos``` because this is just a natural phenomenon that happens every month

it's just a waste of time to cater to their complains when they cant even bother to take 2 screensho```

---

## 25. ```
Claude Suddenly Cant Summarize Previous Cha``` by UUID?
``` {#25-```
claude-suddenly-cant-summarize-previous-cha}

這篇文章的核心討論主題是:**用戶對Claude AI功能無預警變更(尤其是無法存取過往對話歷史以生成摘要)的不滿與困擾**,以及此變更對其工作流程造成的負面影響。

具體要點包括:
1. **功能失效**:原本Claude能透過對話UUID參考歷史內容並生成摘要,但近期突然無法存取過往對話,導致摘要功能失效。
2. **用戶挫折**:作者因依賴此功能管理專案,功能消失後被迫調整工作模式(如頻繁開啟新對話),導致效率下降與時間成本增加。
3. **對Anthropic的不滿**:批評公司未事先公告功能變更,且認為科技公司「暗中收回功能並試圖誤導用戶」(gaslighting)的態度令人反感。
4. **工作流程中斷**:作者原本建立的AI驅動專案管理方法因此受挫,需重新強化自身管理能力填補空缺。

整體聚焦於**AI工具穩定性對用戶信任與工作效能的關鍵影響**,以及科技公司透明度不足的問題。

- **Reddit 連結**: [https://reddit.com/r/ClaudeAI/comments/1jp5xck/claude_suddenly_cant_summarize_previous_chats_by/](https://reddit.com/r/ClaudeAI/comments/1jp5xck/claude_suddenly_cant_summarize_previous_chats_by/)
- **外部連結**: [https://www.reddit.com/r/ClaudeAI/comments/1jp5xck/claude_suddenly_cant_summarize_previous_chats_by/](https://www.reddit.com/r/ClaudeAI/comments/1jp5xck/claude_suddenly_cant_summarize_previous_chats_by/)
- **發布時間**: 2025-04-02 04:27:37

### 內容

When I started working on my project, I asked Claude how best to use Claude to stay organized and it helped me develop a project documentation plan. Core to this effort was Claude summarizing our cha``` in a specific way. When I asked Claude how to maintain that effort if I were to I run into the chat length limit before promoting a given chat for a summary, it told me to reference the UUID (the number after chat/ in the url) and the next Asst would be able to summarize the chat using that reference and my chat summary prompt. This has been working great for the last 5 weeks.

However, starting this weekend, when I ask Claude to do this, it asked me to first summarize what the chat was about. Me : Huh? Thats your job

When I gave it a five word summary (i.e. fixing ongoing form element componen), it spit out a random summary that had nothing to do with what any of my cha had been about. Me : WTF is this?

So I told Claude that I had no idea what it was referring to; and why did it generate randomness. And Claude apologized then told me its not able to access any of my past cha``` so it couldnt help me with this task. Me : Since when?!?

Claude: I can understand youre frustrated and disappointed Me :

Im a month to month subscriber, and for the most part I appreciate how helpful Claude has been for a variety of tasks, but Im just flummoxed as to how Anthropic continues to clawback features at random and without announcement and expec users to just roll with it. Lacking this feature, now Im constantly starting new cha to do very discrete tasks so I can avoid running into the context length/chat limit, but its making it more difficult to maintain a cogent summary of decisions and changes made in a work session, at least doubling the time Im spending on project management.

Cry me a river, I know. Im just bummed because I thought Id figured out the perfect AI-driven project management workflow, and now Ive got to fill the void by stepping up my own project management skills if I ever want to get my project completed. Plus, I hate when tech tries to gaslight me.


### 討論

**評論 1**:

When making a complaint, please

  1. make sure you have chosen the correct flair for the Claude environment that you are using: i.e Web interface (FREE), Web interface (PAID), or Claude API. This information helps others understand your particular situation.
  2. try to include as much information as possible (e.g. prompt and output) so that people can understand the source of your complaint.
  3. be aware that even with the same environment and inpu```, others might have very different outcomes due to Anthropic's testing regime.
  4. be sure to thumbs down unsatisfactory Claude output on Claude.ai. Anthropic representatives tell us they monitor this data regularly.

I am a bot, and this action was performed automatically. Please contact the moderators of this subreddit if you have any questions or concerns.


**評論 2**:

Could you ask Claude to do a summary of your post? Soooooo many words


---

## 26. ```
is it just me or has clause 3.5 gone bonkers today?
``` {#26-```
is-it-just-me-or-has-clause-3-5-gone-bonker}

這篇文章的核心討論主題是:**用戶對Claude AI在編碼過程中突然出現的異常行為(提供未請求的代碼、過度回應等)及其導致專業方案使用時間快速消耗的困擾**。

具體要點包括:
1. **異常行為**:Claude突然違反用戶指令(如主動提供未要求的代碼、在被告知「不要回覆」時仍過度回應)。
2. **效率問題**:相較於過往流暢的體驗,此次異常導致對話效率大幅降低,專業方案的使用時間從1.5-2小時縮短至45分鐘。
3. **資源消耗**:因AI的不必要回應,用戶的付費方案額度被快速耗盡,引發不滿。

整體聚焦於**工具可靠性突變對工作流程與成本的負面影響**。

- **Reddit 連結**: [https://reddit.com/r/ClaudeAI/comments/1jpuaxn/is_it_just_me_or_has_clause_35_gone_bonkers_today/](https://reddit.com/r/ClaudeAI/comments/1jpuaxn/is_it_just_me_or_has_clause_35_gone_bonkers_today/)
- **外部連結**: [https://www.reddit.com/r/ClaudeAI/comments/1jpuaxn/is_it_just_me_or_has_clause_35_gone_bonkers_today/](https://www.reddit.com/r/ClaudeAI/comments/1jpuaxn/is_it_just_me_or_has_clause_35_gone_bonkers_today/)
- **發布時間**: 2025-04-03 01:24:36

### 內容

I've been using claude for coding and it's usually very smooth and the AI does exactly what I ask it to do.

I have 2 pro plans and I go back and forth between 3.7 and 3.5.

For some reason today it keeps trying to give me code I didn't ask for and when i say things like, do not reply it replies with tons of info, or I'll say just provide code for X file and it gives me other files along with the one I asked for.

I used up my professional plan about faster today because of this. I can usually get about 1.5 to 2 hours of chat before my time expires, today it was about 45 minutes worth of chat.


### 討論

**評論 1**:

Giving away unwanted gif``` is 3.7's style. Surprising to hear 3.5 does it too!


---

## 27. ```
Notion MCP tutorial
``` {#27-```
notion-mcp-tutorial
```}

由於我無法直接訪問 YouTube 影片內容,因此無法總結該影片的核心討論主題。不過,您可以根據以下步驟自行分析影片內容:

1. **影片標題與描述**:查看影片的標題和描述,通常會提供主要內容的線索。
2. **關鍵詞與重複概念**:注意影片中反覆出現的詞彙或概念,這些通常是核心主題。
3. **評論區與章節標題**:觀眾評論和影片章節(如果有)可能提煉出重點議題。
4. **演講者或創作者背景**:了解創作者的專業領域,推測其討論方向(例如科技、社會議題等)。

如果您能提供影片的具體內容或關鍵點,我可以幫助進一步分析和總結!

- **Reddit 連結**: [https://reddit.com/r/ClaudeAI/comments/1jpgytq/notion_mcp_tutorial/](https://reddit.com/r/ClaudeAI/comments/1jpgytq/notion_mcp_tutorial/)
- **外部連結**: [https://www.youtube.com/watch?v=Kc_A_Ypah2g](https://www.youtube.com/watch?v=Kc_A_Ypah2g)
- **發布時間**: 2025-04-02 13:10:53

### 內容

連結: [https://www.youtube.com/watch?v=Kc_A_Ypah2g](https://www.youtube.com/watch?v=Kc_A_Ypah2g)

### 討論

無討論內容

---

## 28. ```
Best way to inject a prior chat history seamlessly into a current chat?
``` {#28-```
best-way-to-inject-a-prior-chat-history-sea}

這篇文章的核心討論主題是:
**如何以適當的格式(如XML)遷移部分舊對話到新對話中,並確保AI能正確接續上下文回應(如處理第三條訊息),同時避免重複回應等異常問題。**

具體要點包括:
1. **遷移對話的技術方法**:探討XML標籤(如`<human>`/`<assistant>`)是否適合分段導入歷史對話。
2. **回應品質的穩定性**:目標是讓AI無縫接續新訊息(如第三條),且不降低回答質量。
3. **遇到的技術問題**:使用XML格式時,AI可能重複用戶訊息(如50%機率重複「message 3」),需解決此異常行為。

- **Reddit 連結**: [https://reddit.com/r/ClaudeAI/comments/1jpdt37/best_way_to_inject_a_prior_chat_history/](https://reddit.com/r/ClaudeAI/comments/1jpdt37/best_way_to_inject_a_prior_chat_history/)
- **外部連結**: [https://www.reddit.com/r/ClaudeAI/comments/1jpdt37/best_way_to_inject_a_prior_chat_history/](https://www.reddit.com/r/ClaudeAI/comments/1jpdt37/best_way_to_inject_a_prior_chat_history/)
- **發布時間**: 2025-04-02 10:20:54

### 內容

So I have a prior chat that I want to migrate (not completely) into a fresh chat. What would be the best format or syntax to do that? Claude suggested the XML format:

<human> message 1 </human>

<assistant> response 1 </assistant>

<human> message 2 </human>

<assistant> response 2 </assistant>

<human> message 3 </human>

The goal is to make it respond to message 3 as if the message were following normally in a chat without decrease in quality or bugs.

In fact I experienced bugs with the XML structure above. It replied to message 3, but in 50% of the cases it followed up by repeating message 3 after generating a response 3. Very weird.


### 討論

**評論 1**:

Haven't used the XML format, i usually just write it as

Human: (query)

Assistant: (response)

Human: (query)


**評論 2**:

Easiest? Use a project and teach Claude a memory saving format using the project documen```.

Better way, use MCPs and some other tricks and create your own Jarvis.


**評論 3**:

Are you trying to continue a long chat? I usually just copy the JSON output of the network request, chop off some of the earlier bi``` and tell Claude I want to continue the conversation. Usually works pretty well. But in that case the last message is always the AI response. My latest message would be pasted separately.


**評論 4**:

Use Claude Code. When context is about to run out it automatically saves so it can continue. It's actually excellent at this, best I've seen.

However, you should try to avoid doing this because it does use some context and therefore tokens because of the prior history it's remembering.

Best to use to finish a thought and then issue /clear and start fresh to keep cos``` down.

Claude Code is fantastic.


---

## 29. ```
Worked on a lofi platform
``` {#29-```
worked-on-a-lofi-platform
```}

這篇文章的核心討論主題是:
**作者分享其個人開發的「低傳真(lofi)音樂平台」專案(edenzen.co),並描述其創作動機、學習過程、技術堆疊,以及尋求反饋的請求。**

具體要點包括:
1. **專案介紹**:一個結合專注、放鬆與生產力的lofi音樂平台,主打桌面使用體驗。
2. **個人動機**:出於對lofi音樂的喜愛,並結合自身產品設計背景,跨足開發領域的學習歷程。
3. **技術細節**:使用Next.js、TypeScript、Supabase等工具,並提及Claude AI輔助解決問題。
4. **尋求反饋**:邀請他人試用平台並提供意見,強調專案仍在持續改進中(如未來將優化行動端)。

整體聚焦於「個人side project的開發經驗分享」與「社群互動」。

- **Reddit 連結**: [https://reddit.com/r/ClaudeAI/comments/1jpv4m8/worked_on_a_lofi_platform/](https://reddit.com/r/ClaudeAI/comments/1jpv4m8/worked_on_a_lofi_platform/)
- **外部連結**: [https://www.reddit.com/r/ClaudeAI/comments/1jpv4m8/worked_on_a_lofi_platform/](https://www.reddit.com/r/ClaudeAI/comments/1jpv4m8/worked_on_a_lofi_platform/)
- **發布時間**: 2025-04-03 01:56:16

### 內容

Hey everyone,

I wanted to share a side project Ive been working onedenzen.co. Its a lofi music platform focussed on creating your perfect ambience and digital space.

I started this because Ive always loved the calming vibes of lofi music, and I thought it would be fun to create a space that blends focus, relaxation, and productivity. Im a product designer by trade, but Ive been diving into development, learning as I go, and using Claude heavily to help bridge the gaps in my knowledge.

This has been a huge learning experience, and Id love to hear what you think! Any feedback, feature ideas, or just general though would mean a lot. I far from finished and plenty of things to work on but I would say the base of the web app is complete more or less.

Bear in mind everything above while not complicated at all to a seasoned developer, has been a new experience for me in literally every aspect. My current tech stack is the following

  • Vercel

  • Next.js and Typescript

  • Upstash for rate limiting

  • Cloudflare for CDN

  • Supabase as backend.

I am thankful for my software background which did make it easier to understand and go in when claude keeps looping and saying 'Ah I see the issue'.

Check out edenzen.co and me know what you think!

PS: I``` not optimised for mobile as the platform mainly focuses on desktop and is meant to be used on desktop but eventually will optimise it for mobile.


### 討論

**評論 1**:

Very aestethicly nice, good job. Music is also good, is it from soundcloud?

I also built LoFi radio but with AI generated music :)

https://chillify.me in case you wanna look.


**評論 2**:

Whats the source of the music?


---

## 30. ```
Did Claude get smarter again?
``` {#30-```
did-claude-get-smarter-again-
```}

這段討論的核心主題是:
**使用者觀察到AI模型(Claude 3.7)的回應品質近期出現變化(變得更敏銳、深思熟慮且準確),並詢問這種改進是否為普遍現象,還是個人主觀感受,尤其對比先前其他使用者對模型性能的負面反饋。**

具體要點包括:
1. **效能變化**:使用者感知到模型回應品質的提升(如更準確、更有條理)。
2. **驗證需求**:詢問他人是否也有相同觀察,以確認是系統性更新或個人偏差。
3. **對比爭議**:提及先前社群對模型表現的抱怨,凸顯效能可能存在波動或不穩定。

本質上,這是一個關於**AI模型版本迭代中性能表現的用戶主觀反饋與客觀驗證**的討論。

- **Reddit 連結**: [https://reddit.com/r/ClaudeAI/comments/1jpqet3/did_claude_get_smarter_again/](https://reddit.com/r/ClaudeAI/comments/1jpqet3/did_claude_get_smarter_again/)
- **外部連結**: [https://www.reddit.com/r/ClaudeAI/comments/1jpqet3/did_claude_get_smarter_again/](https://www.reddit.com/r/ClaudeAI/comments/1jpqet3/did_claude_get_smarter_again/)
- **發布時間**: 2025-04-02 22:47:35

### 內容

For the past couple of hours, Claude 3.7 seems noticeably sharper to me responses feel more thoughtful and accurate. Am I the only one noticing this shift, or has something actually changed? Especially since people were complaining about i``` performance earlier this week.


### 討論

**評論 1**:

Anthropic tends to announce builds. Especially since customers rely on consistency.


---

# 總體討論重點

以下是30篇文章的條列式重點總結,並附上對應的錨點連結與逐條細節:

---

### **1. [I blew $417 on Claude Code to build a word game](#anchor_1)**
1. **開發成果**
- 成功用Claude開發文字遊戲LetterLinks,含棋盤設計、計分系統等功能。
2. **AI優點**
- 理解模糊指令(如「棋盤變紫色」)。
- 快速定位錯誤代碼。
3. **AI缺點**
- 上下文限制導致後期代碼記憶困難。
- 修正常引入新錯誤。
- 成本飆升(簡單修改需重新理解代碼)。
4. **經濟性**
- 成本低於真人開發,但心力消耗大。

---

### **2. [Sonnet Computer Use in very Underrated](#anchor_2)**
1. **技術復興**
- Sonnet技術被低估,但潛力大。
2. **應用案例**
- *Apply Hero*、*Vercel*整合瀏覽器代理。
- *BrowserUse*開源商業化成功。
3. **未來趨勢**
- 預測「網頁+電腦代理」應用爆發。

---

### **3. [I regret buying claude for 1 year](#anchor_3)**
1. **版本批評**
- Claude 3.7被評為「fucking shitty」。
2. **用戶挫折**
- 更新後體驗惡化,引發極端情緒(如自傷傾向)。

---

### **4. [Fully Featured AI Coding Agent as MCP Server](#anchor_4)**
1. **工具定位**
- 免費程式碼分析工具Serena,媲美付費方案。
2. **技術創新**
- 使用語言伺服器解析代碼,支援MCP伺服器。
3. **部署選項**
- 可搭配Gemini(需Google Cloud API金鑰)。

---

### **5. [Migrated to Gemini. Sheeesh, the grass is greener.](#anchor_5)**
1. **Claude缺點**
- 上下文長度限制、API過載錯誤。
2. **Gemini優勢**
- 100萬token上下文、更快代碼生成、無訊息限制。
3. **應用成果**
- 順利開發本地化聊天機器人(WolfAI)。

---

### **6. [What the hell happened?](#anchor_6)**
1. **投訴指南**
- 需標註環境類別、提供詳細輸入/輸出。
2. **普遍問題**
- 多名用戶回報系統未更新(附官方狀態頁面)。

---

### **7. [Please be candid; did I just pay $220 for a year of this screensaver?](#anchor_7)**
- 圖片內容無法分析,推測為迷因或諷刺Claude訂閱價值。

---

### **8. [Thanks then! Take care...](#anchor_8)**
1. **技術故障**
- 功能突然失效、無故觸發長度限制錯誤。
2. **用戶協作**
- 釐清問題根源(帳號代幣耗盡或系統端問題)。

---

### **9. [Claude Sonnet is the undisputed champion of OCR](#anchor_9)**
1. **性能排名**
- Claude > 開源模型(Qwen、Mistral) > GPT-4o。
2. **意外發現**
- OpenAI與Anthropic表現差距超出預期。

---

### **10. [Claude 3.7 Sonnet is still the best LLM for frontend development](#anchor_10)**
1. **LLM局限**
- 僅生成靜態模板,缺乏JS邏輯處理。
2. **模型比較**
- Claude 3.7代碼格式不穩,Gemini 2.5 Pro更強但輸出混亂。

---

(因篇幅限制,以下簡化條列標題,完整細節請參照錨點連結)

### **11-30. 快速摘要**
- **#11** 從Claude轉向Gemini的情感糾結與效能比較。
- **#12** Claude Code效能佳但成本過高,盼降價。
- **#13** 遊戲開發協作與發布過程的非正式討論。
- **#14** 質疑Claude摘要壓縮機制犧牲上下文品質。
-