From 4301500ee3cdce9dd547795e42ff16b281dba202 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E6=9F=B3=E7=A5=9E?= <3162475700@qq.com> Date: Sun, 3 Nov 2024 23:15:01 +0800 Subject: [PATCH] =?UTF-8?q?=F0=9F=98=8B=E4=BB=8E150=E5=BC=80=E5=A7=8B?= =?UTF-8?q?=E5=88=A4=E6=96=AD=EF=BC=8C=E8=BF=9B=E4=B8=80=E6=AD=A5=E7=BC=A9?= =?UTF-8?q?=E5=B0=8F=E6=96=87=E4=BB=B6=E5=A4=A7=E5=B0=8F?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- friend_circle_lite/get_info.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/friend_circle_lite/get_info.py b/friend_circle_lite/get_info.py index 2133690..7a128b9 100644 --- a/friend_circle_lite/get_info.py +++ b/friend_circle_lite/get_info.py @@ -420,8 +420,8 @@ def deal_with_large_data(result): # 检查文章数量是否大于 150 if len(article_data) > 150: print("数据量较大,开始进行处理···") - # 获取前 200 篇文章的作者集合 - first_200_authors = {article["author"] for article in article_data[:200]} + # 获取前 150 篇文章的作者集合 + first_200_authors = {article["author"] for article in article_data[:150]} # 从第151篇开始过滤,只保留前150篇出现过的作者的文章 filtered_articles = article_data[:150] + [