diff --git a/models/database.db b/models/database.db index 4a1b9d1..cdcb63b 100644 Binary files a/models/database.db and b/models/database.db differ diff --git a/telegram/8619211027341.session b/telegram/8619211027341.session index 93e2aa7..6972a42 100644 Binary files a/telegram/8619211027341.session and b/telegram/8619211027341.session differ diff --git a/telegram/bot_session.session b/telegram/bot_session.session index 6505eaa..ece3268 100644 Binary files a/telegram/bot_session.session and b/telegram/bot_session.session differ diff --git a/telegram/sign.db b/telegram/sign.db index 0c3c0ef..d15468b 100644 Binary files a/telegram/sign.db and b/telegram/sign.db differ diff --git a/weex/自动化抓取数据.py b/weex/自动化抓取数据.py index 1bf8e70..e12c7f8 100644 --- a/weex/自动化抓取数据.py +++ b/weex/自动化抓取数据.py @@ -6,7 +6,7 @@ from models.weex import Weex1, Weex1Hour, Weex15, Weex30 if __name__ == '__main__': - bit_port = openBrowser(id="f2320f57e24c45529a009e1541e25961") + bit_port = openBrowser(id="86837a981aba4576be6916a0ef6ad785") # response = requests.post( # f"http://127.0.0.1:50326/api/browser/start", diff --git a/推特/main.py b/推特/main.py index 5a16998..83114da 100644 --- a/推特/main.py +++ b/推特/main.py @@ -589,9 +589,9 @@ Win–win benefits — don’t miss out! 🚀 urls = [ "https://x.com/Websea_MY/status/2004081342364123146", - # "https://x.com/Websea_MY/status/2003669526492405904", - # # "https://x.com/Websea_MY/status/2001119115575222779", - # # "https://x.com/Websea_MY/status/2001180999443734677" + "https://x.com/Websea_MY/status/2003669526492405904", + "https://x.com/Websea_MY/status/2001119115575222779", + "https://x.com/Websea_MY/status/2001180999443734677" ] tests = [ "O seguro de futuros da Websea ultrapassou oficialmente a marca de 400 rodadas de airdrops! 🚀 Contacte TG@webseatom 85% de reembolso", @@ -743,51 +743,51 @@ def run_work(x_token_info, xstart_info): pass # hub.to_do_tui() # 发推 - # hub.回复() # 发推 + hub.回复() # 发推 - hub.FollowTwitterAccount() # 关注 + # hub.FollowTwitterAccount() # 关注 if __name__ == '__main__': - # fz_datas = get_group_lists() + fz_datas = get_group_lists() + + for bit_data in get_browser_lists_Browser(id=fz_datas['推特']): + print(bit_data) + xstart_info, start = Xstart.get_or_create( + bit_id=bit_data["id"], + ) + if xstart_info.x_id: + # continue + + hun_web = Hub_Web(xstart_info=xstart_info) + + # hun_web.action() + + threading.Thread(target=hun_web.action).start() + time.sleep(random.randint(15, 60)) + + # # 同时运行 + # max_threads = 5 + # delay_between_start = 30 # 每次启动线程之间的延迟时间(秒) # - # for bit_data in get_browser_lists_Browser(id=fz_datas['推特']): - # print(bit_data) - # xstart_info, start = Xstart.get_or_create( - # bit_id=bit_data["id"], - # ) - # if xstart_info.x_id: - # # continue + # with ThreadPoolExecutor(max_workers=max_threads) as executor: # - # hun_web = Hub_Web(xstart_info=xstart_info) + # # for x_token_info in XToken.select(): # - # # hun_web.action() + # # 查询数据并转换为列表 + # random_x_infos = list(XToken.select()) + # # 直接对原列表进行打乱操作 + # random.shuffle(random_x_infos) # - # threading.Thread(target=hun_web.action).start() - # time.sleep(random.randint(15, 60)) - - # 同时运行 - max_threads = 5 - delay_between_start = 30 # 每次启动线程之间的延迟时间(秒) - - with ThreadPoolExecutor(max_workers=max_threads) as executor: - - # for x_token_info in XToken.select(): - - # 查询数据并转换为列表 - random_x_infos = list(XToken.select()) - # 直接对原列表进行打乱操作 - random.shuffle(random_x_infos) - - # 遍历打乱顺序后的列表 - for x_token_info in random_x_infos: - xstart_info, start = Xstart.get_or_create( - x_id=x_token_info.id, - ) - - # if xstart_info.start: - # continue - - executor.submit(run_work, x_token_info, xstart_info) - # time.sleep(random.randint(60, 150)) - time.sleep(delay_between_start) + # # 遍历打乱顺序后的列表 + # for x_token_info in random_x_infos: + # xstart_info, start = Xstart.get_or_create( + # x_id=x_token_info.id, + # ) + # + # # if xstart_info.start: + # # continue + # + # executor.submit(run_work, x_token_info, xstart_info) + # # time.sleep(random.randint(60, 150)) + # time.sleep(delay_between_start)