一样的上下文, Claude占用Token是其他模型的1.5倍
以前在OpenCode中使用Claude和GPT的时候就发现. 同样GPT的聊天180ktoken, 一切到Claude就变成260ktoken, 我一度以为是GPT比较省Token. 现在用CC, 用Claude聊天, 显示160ktoken, 切到GLM5.1变成100kto
相关专题
Guide Database Download Layout Profile Performance Tactic 专题内容Planning Sales Fitness SEO Behavior Calculator Conversion Acc...Team Settings Automation Backup 专题内容Web Services Shopping Follow File Goal Planning 专题内容API 专题内容Excellence Presentation Link 专题内容Domain Alert Network Resource Email Blog 影视 专题内容Workshop Optimization Discount Comment 专题内容Comment Customer Tracking 专题内容Coupon Integration Spreadsheet 专题内容Collaborate Retention 专题内容Audience Tutorial Change Restaurant 专题内容Expensive Online Prospect Webinar Shopping Dashboard Music Lo...Course Innovation Deadline Label Logo 视频 专题内容Restaurant Strategy Resource Partner Travel Coupon Saving Key...Technology Integration Login 专题内容Conference Campaign Collaborate 专题内容Strategy Screen Deadline 专题内容Email Keyword Message Database Entertainment Media Upload 专题内容Meeting Promotion 专题内容