Przeglądaj źródła

Merge branch 'master' of http://git.choozmo.com:3000/choozmo/kw_tools

Jason 2 lat temu
rodzic
commit
60795a5218

+ 11 - 11
SEO/clickbot_100.py

@@ -43,26 +43,26 @@ def restart_browser():
 
 def process_one():
     db = dataset.connect('mysql://choozmo:pAssw0rd@db.ptt.cx:3306/seo?charset=utf8mb4')
-    lst=['台中 留學代辦']
+    lst=[]
     table=db['google_rank']
-    # cursor = db.query('select term,domain from seo.selected_kw where client="CLIQ露營椅"')
+    # cursor = db.query('select term,domain from seo.selected_kw where client="護全"')
     # cursor = db.query('select term,url from seo.sns_kw where client="loginheart"')
     # cursor = db.query('select term from seo.selected_kw where client="鑫富"')
-    # cursor = db.query('select term from seo.selected_kw where id between 1902 and 1923')
+    cursor = db.query('select term,domain from seo.selected_kw where id between 2146 and 2155')
     # cursor=db.query('select term from selected_kw and term not in (SELECT distinct(keyword) FROM ig_tags.save_result where url like "%beastparadise.net%" and datediff(now(),dt)=0)')
-    # for c in cursor:
-    #     lst.append([c['term'],c['domain']])
+    for c in cursor:
+        lst.append([c['term'],c['domain']])
 
-    domain = 'hsinfei'
+    # domain = 'wandatw.com'
     for i in lst:
         print(i)
         driver=restart_browser()
-        escaped_search_term=urllib.parse.quote(i)
+        escaped_search_term=urllib.parse.quote(i[0])
         googleurl = 'https://www.google.com/search?q={}&num={}&hl={}'.format(escaped_search_term, 100,'zh-TW')
         driver.get(googleurl)
         time.sleep(60)
         print(driver.current_url)
-        # driver.save_screenshot('C:\/Users\/s1301\/Pictures\/Saved Pictures\/CLIQ\/'+i[0]+'.png')
+        driver.save_screenshot('C:\/Users\/s1301\/Pictures\/Saved Pictures\/護全\/'+i[0]+'.png')
         df=pd.DataFrame()
         elmts=driver.find_elements(By.XPATH,"//div[@class='yuRUbf']/a")
         print('結果數量',len(elmts))
@@ -72,8 +72,8 @@ def process_one():
         for elmt in elmts:
             try:
                 href=elmt.get_attribute('href')
-                if domain in href:
-                    table.insert({'title':elmt.text,'url':href,'keyword':i,'dt':datetime.datetime.now(),'ranking':cnt})
+                if i[1] in href:
+                    table.insert({'title':elmt.text,'url':href,'keyword':i[0],'dt':datetime.datetime.now(),'ranking':cnt})
                     print(href)
                     print(elmt.text)
                 # datadict['搜尋詞'].append(term)
@@ -92,7 +92,7 @@ def process_one():
         df['結果網址']=datadict['結果網址']
         df['結果名次']=datadict['結果名次']
 
-        # df.to_excel('C:\/Users\/s1301\/Pictures\/Saved Pictures\/鑫富\/'+term+".xls")
+        df.to_excel('C:\/Users\/s1301\/Pictures\/Saved Pictures\/護全\/'+i[0]+".xls")
 
         driver.quit()
         print('等待')

+ 3 - 8
SEO/clickbot_100_one.py

@@ -41,8 +41,8 @@ def restart_browser():
     return driver
 
 def process_one():
-    lst=['雙響泡','雙響砲','双響泡']
-    date='1214'
+    lst=['龔國權']
+    date='1223'
     for term in lst:
         driver=restart_browser()
         escaped_search_term=urllib.parse.quote(term)
@@ -87,10 +87,5 @@ def process_one():
 process_one()
 
 
-os.system('docker container stop tiny8')
-time.sleep(3)
-os.system('docker container rm tiny8')
-time.sleep(3)
-os.system('docker run -d -p 9924:9222 --rm --shm-size="900m" --name tiny1 chromedp/headless-shell')
-time.sleep(3)
+
 

+ 13 - 9
SEO/csv_to_sql.py

@@ -9,32 +9,36 @@ pymysql.install_as_MySQLdb()
 db = dataset.connect('mysql://choozmo:pAssw0rd@db.ptt.cx:3306/seo?charset=utf8mb4')
 table=db['seo_jobs']
 # table=db['selected_kw']
-client='泡麵'
-domain=''
+# table=db['sns_kw']
+client='信義房屋'
+domain='www.hujiai-international.com'
 # for index,row in df.iterrows():
 # with open("C:\/Users\/s1301\/Documents\/新飛國際遊學SEO - 關鍵字12.08.csv") as f:
 #     data_all = f.readlines()
 # print(data_all)
 f = open("C:\/Users\/s1301\/Documents\/456.csv",encoding='utf-8')
 # df = pd.read_csv(f,header=None, names=['kw', 'url'])
-df = pd.read_csv(f,header=None, names=['prefix', 'positive','rnd','URL'])
+# df = pd.read_csv(f,header=None, names=['kw'])
+df = pd.read_csv(f,header=None, names=['prefix', 'positive','rnd','domain'])
 df=df.fillna('')
 # print(df)
-url='cliqproducts.com.tw'
-
+# url='fflc.tw'
+lst=[]
 for index,row in df.iterrows():
     # print(row['kw'],url)
+    # print(row['kw'])
+    # lst.append(row['kw'])
     prefix='"'+row['prefix']+'"'
     positive='"'+row['positive']+'"'
-    domain='"'+row['URL']+'"'
+    domain='"'+row['domain']+'"'
     rnd='"'+str(row['rnd'])+'"'
     postfix='""'
-    data = f'"prefix":{prefix},"postfix":{postfix},"positive":[{positive}],"domain":[{domain}],"rnd":[{rnd}]'
+    data = f'"prefix":{prefix},"domain":[{domain}],"postfix":{postfix},"positive":[{positive}],"rnd":[{rnd}]'
     json='{'+data+'}'
     print(json)
     table.insert({'cust':client,'plan':'形象SEO','json':json})
-    # table.insert({'term':row['kw'],'client':client,'domain':url})
-
+    # table.insert({'term':row['kw'],'client':client,'domain':domain})
+print(lst)
 {"prefix": "護佳", "postfix": "", "domain": ["www.hujiai-international.com"], "positive": ["細胞食物","紅蔘"], "rnd": [""]}
 
 db.close()

+ 1 - 1
SEO/google_rank.py

@@ -91,7 +91,7 @@ def run_once(q):
 #    sleepint=random.randint(320,520)
 #    time.sleep(sleepint)
 
-lst = ['龔國權','信義房屋','護佳國際','信義房仲','信義 房屋','信義 房仲','雙響泡']
+lst = ['信義房仲','信義 房屋','信義 房仲','雙響泡']
 for i in lst:
     run_once(i)
 

+ 5 - 2
SEO/mysql_to_redis.py

@@ -8,11 +8,14 @@ import json
 db = dataset.connect('mysql://choozmo:pAssw0rd@db.ptt.cx:3306/seo?charset=utf8mb4')
 table=db['selected_kw']
 lst = []
-cur = db.query('select term,url from seo.sns_kw where client="loginheart"')
+cur = db.query('select term,url from seo.sns_kw where client="理茶"')
 for c in cur:
     lst.append([c['term'],c['url']])
 
+
 print(lst)
 # r = redis.Redis(host='db.ptt.cx', port=6379, db=1,password='choozmo9')
 # r.set(lst)
-['真皮迷你包', 'https://www.zeczec.com/projects/loginheart'], ['小廢包 品牌', 'https://www.zeczec.com/projects/loginheart'], ['電子支付包包', 'https://www.zeczec.com/projects/loginheart'], ['感應支付 小包包', 'https://www.zeczec.com/projects/loginheart'], ['一卡通小包', 'https://www.zeczec.com/projects/loginheart'], ['迷你包 推薦', 'https://www.zeczec.com/projects/loginheart'], ['科技 小廢包', 'https://www.zeczec.com/projects/loginheart'], ['小配方包', 'https://www.zeczec.com/projects/loginheart'], ['精品包包', 'https://www.zeczec.com/projects/loginheart'], ['時尚科技 台灣 包包品牌', 'https://www.zeczec.com/projects/loginheart']
+['中 美 街 飲料', 'https://www.daisyyohoho.com/richa-drinks/'], ['中 美 街 飲料', 'https://www.walkerland.com.tw/article/view/340526'], ['中 美 街 飲料', 'https://www.facebook.com/Richa.drinks/'], ['中 美 街 飲料店', 'https://www.walkerland.com.tw/article/view/340526'], ['中 美 街 飲料 店', 'https://chiaoda.com/richa/'], ['勤美飲料店', 'https://playqueen888.com/tourall.php?twId=772&twlistId=1'], ['台中必喝飲料', 'https://playqueen888.com/tourall.php?twId=772&twlistId=1'], ['勤美 飲料 內用', 'https://www.walkerland.com.tw/article/view/340526'], ['精誠 路 飲料 店', 'https://chiaoda.com/richa/'], ['中 美 街 飲料 店推薦', 'https://zh-tw.facebook.com/Richa.drinks/'], ['台中中美街飲料', 'https://www.daisyyohoho.com/richa-drinks/'], ['中美街飲料推薦', 'https://www.daisyyohoho.com/richa-drinks/'], ['內用的手搖飲料店', 'https://taiwan17go.com/richa/'], ['台中限定飲料店', 'https://taiwan17go.com/richa/'], ['台中西區飲料推薦', 'https://chiaoda.com/richa/'], ['台中內用飲料店', 'https://chiaoda.com/richa/'], ['精誠路飲料', 'https://chiaoda.com/richa/'], ['台中中美街飲料', 'https://www.facebook.com/Richa.drinks/'], ['中美街 手搖飲', 'https://www.walkerland.com.tw/article/view/340526'], ['中美街 手搖飲', 'https://wandatw.com/richa/'], ['中美街手搖', 'https://www.daisyyohoho.com/richa-drinks/'], ['精誠路手搖', 'https://chiaoda.com/richa/']
+
+#####用記事本替換#######

+ 2 - 2
SEO/ranking_report.py

@@ -3,8 +3,8 @@ import dataset
 import pymysql
 pymysql.install_as_MySQLdb()
 
-name='宏康'
-date = '1215'
+name='plantA'
+date = '1222'
 
 
 def day():

+ 10 - 4
SEO/ranking_world.py

@@ -134,6 +134,7 @@ def run_once(jsobj):
         driver.get('https://www.google.com?num=100&lr=lang_en')
 
         time.sleep(30)
+        driver.get('https://www.google.com?num=100&lr=lang_en')
         print(driver.current_url)
         elmt = driver.find_element(By.XPATH, "//input[@name='q']")
         time.sleep(1)
@@ -145,8 +146,8 @@ def run_once(jsobj):
         elmts = driver.find_elements(By.XPATH, "//div[@class='yuRUbf']/a")
         numresults = len(elmts)
         print('搜尋結果數量', numresults)
-        time.sleep(30)
-        driver.save_screenshot('C:\/Users\/s1301\/Pictures\/Saved Pictures\/' +date +fname + ".png")
+        time.sleep(20)
+        driver.save_screenshot('C:\/Users\/s1301\/Pictures\/Saved Pictures\/angelo koo\/' +date +fname + ".png")
         if numresults == 0:
             send_msg('stop working...')
             sys.exit()
@@ -218,7 +219,7 @@ def run_once(jsobj):
         df['結果名次'] = datadict['結果名次']
         df['結果說明'] = datadict['結果說明']
 
-        df.to_excel('C:\/Users\/s1301\/Pictures\/Saved Pictures\/'+date+fname+".xls")
+        df.to_excel('C:\/Users\/s1301\/Pictures\/Saved Pictures\/angelo koo\/'+date+fname+".xls")
 
 
 
@@ -233,6 +234,11 @@ def run_once(jsobj):
     sys.exit()
 
 
-run_once({'kw':'angelo koo','fname':'ny','date':'1215_2'})
+run_once({'kw':'angelo koo','fname':'chicago','date':'1222'})
 
+####手動截圖:須按右下角的設定選擇區域######
+ny="https://www.google.com/search?q=angelo+koo&hl=en&gl=us&num=100&uule=w+CAIQICIWTmV3IFlvcmssVW5pdGVkIFN0YXRlcw&gws_rd=cr"
+wc="https://www.google.com/search?q=angelo+koo&hl=en&gl=us&uule=w+CAIQICIYV2FzaGluZ3RvbixVbml0ZWQgU3RhdGVz&gws_rd=cr"
+miami="https://www.google.com/search?q=angelo+koo&hl=en&gl=us&uule=w+CAIQICIbTWlhbWksRmxvcmlkYSxVbml0ZWQgU3RhdGVz&gws_rd=cr"
+chicago="https://www.google.com/search?q=angelo+koo&hl=en&gl=us&uule=w+CAIQICIeQ2hpY2FnbyxJbGxpbm9pcyxVbml0ZWQgU3RhdGVz&gws_rd=cr"
 

+ 217 - 0
SEO/ranking_world_2.py

@@ -0,0 +1,217 @@
+# import redis
+import time
+import traceback
+# import json
+from selenium import webdriver
+from selenium.webdriver.common.desired_capabilities import DesiredCapabilities
+import time
+import os
+from selenium.webdriver.support.ui import WebDriverWait
+from selenium.webdriver.common.by import By
+from selenium.webdriver.support import expected_conditions as EC
+import dataset
+from selenium.webdriver.common.keys import Keys
+import json
+import random
+import time
+import redis
+import sys
+import codecs
+import pandas as pd
+import random
+import os
+import time
+import datetime
+from selenium.webdriver.chrome.service import Service
+import dataset
+import pymysql
+
+pymysql.install_as_MySQLdb()
+
+from userAgentRandomizer import userAgents
+import requests
+
+driver = None
+
+db = dataset.connect('mysql://choozmo:pAssw0rd@db.ptt.cx:3306/seo?charset=utf8mb4')
+
+headers = {
+    "Authorization": "Bearer " + "6SDULL1Ebklduc6TFxa97AFto5Sj21kyJ30CxiLiSoi",
+    "Content-Type": "application/x-www-form-urlencoded"
+}
+
+
+def send_msg(kw):
+    params = {"message": kw}
+    r = requests.post("https://notify-api.line.me/api/notify", headers=headers, params=params)
+
+
+def re_get_webdriver():
+    global port
+    global driver
+    result = []
+    if driver is not None:
+        print('closing....')
+        driver.quit()
+        os.system('killall chrome')
+        print('quit....')
+        driver = None
+    try:
+        ua = userAgents()
+
+        user_agent = ua.random()
+
+        options = webdriver.ChromeOptions()
+        options.add_argument("--no-sandbox")
+        options.add_argument("--disable-dev-shm-usage")
+        # options.add_argument("--headless")
+        print(user_agent)
+        options.add_experimental_option('prefs', {'intl.accept_languages': 'en,en_US'})
+        options.add_argument("--incognito")
+        driver = None
+        try:
+            if os.name == 'nt':
+                driver = webdriver.Chrome(options=options)
+
+            else:
+                driver = webdriver.Chrome(executable_path=r'C:\Users\Administrator\Downloads\chromedriver_108\chromedriver', options=options)
+        except:
+            traceback.print_exc()
+            return
+        driver.delete_all_cookies()
+        driver.set_window_size(950, 20000)
+        return
+    except:
+        traceback.print_exc()
+        driver = None
+        return None
+
+
+def scrolling(driver, pgnum):
+    ub = driver.find_element("css selector",'body')
+    for i in range(pgnum):
+        ub.send_keys(Keys.PAGE_DOWN)
+        if pgnum > 1:
+            time.sleep(0.3)
+
+
+def run_once(jsobj):
+    table = db['nda_log']
+
+    print(jsobj)
+    global driver
+
+    if driver is None:
+        time.sleep(8)
+        re_get_webdriver()
+    if driver is None:
+        return
+    try:
+        kw = jsobj['kw']
+        url = jsobj['url']
+
+        # if jsobj.get('domain') is None:
+        #     exclude = jsobj['exclude']
+        #     domain = None
+        # else:
+        #     domain = jsobj['domain']
+        #     exclude = None
+
+        driver.get(url)
+        time.sleep(6)
+        scrolling(driver,10)
+        time.sleep(20)
+        elmts = driver.find_elements(By.XPATH, "//div[@class='yuRUbf']/a")
+        numresults = len(elmts)
+        print('搜尋結果數量', numresults)
+        time.sleep(20)
+        # driver.save_screenshot('C:\/Users\/s1301\/Pictures\/Saved Pictures\/angelo koo\/' +date +fname + ".png")
+        if numresults == 0:
+            send_msg('stop working...')
+            sys.exit()
+        datadict = {'搜尋詞': [], '結果標題': [], '結果網址': [], '結果名次': [], '結果說明': []}
+        df = pd.DataFrame()
+
+        idx = 1
+        found = False
+        test_lst = []
+        txt_dict = {}
+        for elmt in elmts:
+            href = elmt.get_attribute('href')
+            txt = elmt.text
+            desc = None
+
+            if domain in href:
+                print(href)
+                print("ranking", idx)
+
+                webdriver.ActionChains(driver).move_to_element(elmt).perform()
+                # elmt.click()
+                webdriver.ActionChains(driver).move_to_element(elmt).click().perform()
+                print('clicked....')
+                table.insert({'ranking': idx, 'kw': kw, 'results': numresults, 'url': href, 'title': txt,
+                              'dt': datetime.datetime.now(), 'client': 'i5'})
+                time.sleep(6)
+                db.close()
+
+            idx += 1
+
+
+        # if exclude is not None:
+        #     print('exclude')
+        #     elmt = random.choice(test_lst[5:])
+        #     print(elmt)
+        #     print(txt_dict[elmt])
+        #
+        #     webdriver.ActionChains(driver).move_to_element(elmt).perform()
+        #     #            elmt.click()
+        #     webdriver.ActionChains(driver).move_to_element(elmt).click().perform()
+        #     time.sleep(5)
+        #
+        # if not found:
+        #     table.insert(
+        #         {'ranking': -1, 'kw': kw, 'results': numresults, 'url': '', 'title': '未收錄', 'descrption': desc})
+
+        df['搜尋詞'] = datadict['搜尋詞']
+        df['結果標題'] = datadict['結果標題']
+        df['結果網址'] = datadict['結果網址']
+        df['結果名次'] = datadict['結果名次']
+        df['結果說明'] = datadict['結果說明']
+
+        # df.to_excel('C:\/Users\/s1301\/Pictures\/Saved Pictures\/angelo koo\/'+date+fname+".xls")
+
+
+
+
+    except:
+        print('exception')
+        traceback.print_exc()
+    #    time.sleep(9999)
+    #    driver.save_screenshot('c:/tmp/seo/'+kw+".png")
+
+    driver.quit()
+    sys.exit()
+
+
+
+city_lst = ["w+CAIQICIWTmV3IFlvcmssVW5pdGVkIFN0YXRlcw&gws_rd=cr",
+            "w+CAIQICIYV2FzaGluZ3RvbixVbml0ZWQgU3RhdGVz&gws_rd=cr",
+            "w+CAIQICIbTWlhbWksRmxvcmlkYSxVbml0ZWQgU3RhdGVz&gws_rd=cr",
+            "w+CAIQICIeQ2hpY2FnbyxJbGxpbm9pcyxVbml0ZWQgU3RhdGVz&gws_rd=cr"]
+
+
+cursor=db.query('select json from seo_jobs where cust="美東" and plan="形象SEO" order by rand() limit 1')
+for c in cursor:
+    js=json.loads(c['json'])
+    prefix=js['prefix']
+    postfix=js['postfix']
+    domain=js['domain'][0]
+    positive=js['positive']
+    rnd=js['rnd']
+
+kw1=random.choice(positive)
+kw2=random.choice(rnd)
+kw=prefix+" "+kw1+" "+kw2
+city = random.choice(city_lst)
+url = f"https://www.google.com/search?q={kw}&hl=en&gl=us&num=100&uule={city}&gws_rd=cr"
+run_once({'kw':kw,'url':url})

+ 1 - 1
SEO/sns_rank_report.py

@@ -4,7 +4,7 @@ import pymysql
 pymysql.install_as_MySQLdb()
 
 name='superlink'
-date = '1212'
+date = '1219'
 
 
 def day():

+ 94 - 80
monitor/GA_all.py

@@ -8,16 +8,16 @@ import pandas as pd
 import random
 import urllib
 import json
-import gspread
+
 import datetime
-from gspread_pandas import Spread, Client
-from oauth2client.service_account import ServiceAccountCredentials
+
 import os
 import threading
 from googleapiclient.discovery import build
 from oauth2client.service_account import ServiceAccountCredentials
 import dataset
 import pymysql
+
 pymysql.install_as_MySQLdb()
 from datetime import datetime
 import platform
@@ -26,33 +26,34 @@ db = dataset.connect('mysql://choozmo:pAssw0rd@db.ptt.cx:3306/hhh?charset=utf8mb
 db.query('delete from ga_pagepath')
 db.begin()
 
-table=db['ga_pagepath']
+table = db['ga_pagepath']
 SCOPES = ['https://www.googleapis.com/auth/analytics.readonly']
 platform_now = platform.system().lower()
 
 # KEY_FILE_LOCATION = 'c:\gitlab\kw_tools\monitor\corded-velocity-301807-a3e3d5420aba.json'
 KEY_FILE_LOCATION = '/Users/zooeytsai/Downloads/corded-velocity-301807-a3e3d5420aba.json'
 
-
 # line notify header
 headers = {
-        "Authorization": "Bearer " + "QCAM5upFjeBVp54PGqT4eMZSXPU0y4vYk1e1CoASa2P",
-        "Content-Type": "application/x-www-form-urlencoded"
+    "Authorization": "Bearer " + "QCAM5upFjeBVp54PGqT4eMZSXPU0y4vYk1e1CoASa2P",
+    "Content-Type": "application/x-www-form-urlencoded"
 }
 
 com_table = []
+
+
 def creat_table():
-  for i in range(0,24):
-    com_table.append([i,6000])
-  # com_table.append([24,70000])
+    for i in range(0, 24):
+        com_table.append([i, 6000])
+    # com_table.append([24,70000])
 
 
-def send_msg_pg(pv):
-  # line notify send message
-    current_time = datetime.now().strftime('%Y-%m-%d %H:%M:%S') #現在時間
-    hour = datetime.now().strftime('%H') 
+def send_msg_pg(viewid, pv):
+    # line notify send message
+    current_time = datetime.now().strftime('%Y-%m-%d %H:%M:%S')  # 現在時間
+    hour = datetime.now().strftime('%H')
     # 判斷是否達標
-    complet="否"
+    complet = "否"
     # for i in range(0,25):
     #   if int(hour)+1==com_table[i][0]:
     #     print(i)
@@ -61,104 +62,117 @@ def send_msg_pg(pv):
     #   elif int(hour) == 24:
     #     if int(kw) > 70000 :
     #       complet="是"
-    print('網頁瀏覽量',pv)
-    if int(pv) > 6000:
-      complet = "是"
-    params = {"message": "\n現在時間: " + current_time + "\n當前pageViews: "+pv + "\n是否達標: " + complet}
+    print('網頁瀏覽量', pv)
+    if viewid == '208868237':
+        if int(pv) > 2000:
+            complet = "是"
+        else:
+            complet = '否'
+    else:
+        if int(pv) > 2300:
+            complet = "是"
+        else:
+            complet = "否"
+
+    params = {"message": "\n現在時間: " + current_time + "\n當前pageViews: " + pv + "\n是否達標: " + complet}
     print(params)
     return params
-    r = requests.post("https://notify-api.line.me/api/notify",headers=headers, params=params)
+    r = requests.post("https://notify-api.line.me/api/notify", headers=headers, params=params)
     print(r)
 
 
 def initialize_analyticsreporting(key_file):
-  """Initializes an Analytics Reporting API V4 service object.
+    """Initializes an Analytics Reporting API V4 service object.
 
   Returns:
     An authorized Analytics Reporting API V4 service object.
   """
-  credentials = ServiceAccountCredentials.from_json_keyfile_name(
-      key_file, SCOPES)
+    credentials = ServiceAccountCredentials.from_json_keyfile_name(
+        key_file, SCOPES)
 
-  # Build the service object.
-  analytics = build('analyticsreporting', 'v4', credentials=credentials)
+    # Build the service object.
+    analytics = build('analyticsreporting', 'v4', credentials=credentials)
 
-  return analytics
+    return analytics
 
 
-def get_report(analytics,body):
-  """Queries the Analytics Reporting API V4.
+def get_report(analytics, body):
+    """Queries the Analytics Reporting API V4.
 
   Args:
     analytics: An authorized Analytics Reporting API V4 service object.
   Returns:
     The Analytics Reporting API V4 response.
   """
-  return analytics.reports().batchGet(
-      body={
-        'reportRequests':body
-      }
-  ).execute()
+    return analytics.reports().batchGet(
+        body={
+            'reportRequests': body
+        }
+    ).execute()
 
 
 def print_response(response):
-  """Parses and prints the Analytics Reporting API V4 response.
+    """Parses and prints the Analytics Reporting API V4 response.
 
   Args:
     response: An Analytics Reporting API V4 response.
   """
-  result=[]
-  for report in response.get('reports', []):
-    columnHeader = report.get('columnHeader', {})
-    dimensionHeaders = columnHeader.get('dimensions', [])
-    metricHeaders = columnHeader.get('metricHeader', {}).get('metricHeaderEntries', [])
-
-    for row in report.get('data', {}).get('rows', []):
-      dimensions = row.get('dimensions', [])
-      dateRangeValues = row.get('metrics', [])
-      ga_dict={}
-
-      for header, dimension in zip(dimensionHeaders, dimensions):
-#        print(header + ': ', dimension)
-        ga_dict[header]=dimension
-      
-      for i, values in enumerate(dateRangeValues):
-#        print('Date range:', str(i))
-        for metricHeader, value in zip(metricHeaders, values.get('values')):
-          ga_dict[metricHeader.get('name')]=value
-#          print(metricHeader.get('name') + ':', value)
-        result.append(ga_dict)
-    return result
+    result = []
+    for report in response.get('reports', []):
+        columnHeader = report.get('columnHeader', {})
+        dimensionHeaders = columnHeader.get('dimensions', [])
+        metricHeaders = columnHeader.get('metricHeader', {}).get('metricHeaderEntries', [])
+
+        for row in report.get('data', {}).get('rows', []):
+            dimensions = row.get('dimensions', [])
+            dateRangeValues = row.get('metrics', [])
+            ga_dict = {}
+
+            for header, dimension in zip(dimensionHeaders, dimensions):
+                #        print(header + ': ', dimension)
+                ga_dict[header] = dimension
+
+            for i, values in enumerate(dateRangeValues):
+                #        print('Date range:', str(i))
+                for metricHeader, value in zip(metricHeaders, values.get('values')):
+                    ga_dict[metricHeader.get('name')] = value
+                #          print(metricHeader.get('name') + ':', value)
+                result.append(ga_dict)
+        return result
+
+
 #        print(ga_dict)
 
 
-def main(viewid,key_file):
-  creat_table()
-  analytics = initialize_analyticsreporting(key_file)
-  current_time = datetime.now().strftime('%Y-%m-%d') #現在時間
-  body=[{ 'viewId': viewid,
-  'dateRanges': [{'startDate': current_time, 'endDate': current_time}],
+def main(viewid, key_file):
+    creat_table()
+    analytics = initialize_analyticsreporting(key_file)
+    current_time = datetime.now().strftime('%Y-%m-%d')  # 現在時間
+    body = [{'viewId': viewid,
+             'dateRanges': [{'startDate': current_time, 'endDate': current_time}],
 
-  'metrics': [{'expression': 'ga:users'},{'expression': 'ga:newusers'},{'expression': 'ga:sessions'},{'expression': 'ga:pageviews'},{'expression': 'ga:bounceRate'},{'expression': 'ga:pageviewsPerSession'}],
-#  'dimensions': [{'name': 'ga:pagePath'}],
-#  'orderBys':[{"fieldName": "ga:pageviews", "sortOrder": "DESCENDING"}],
-  'pageSize': '100'
-  }]
+             'metrics': [{'expression': 'ga:users'}, {'expression': 'ga:newusers'}, {'expression': 'ga:sessions'},
+                         {'expression': 'ga:pageviews'}, {'expression': 'ga:bounceRate'},
+                         {'expression': 'ga:pageviewsPerSession'}],
+             #  'dimensions': [{'name': 'ga:pagePath'}],
+             #  'orderBys':[{"fieldName": "ga:pageviews", "sortOrder": "DESCENDING"}],
+             'pageSize': '100'
+             }]
 
-  response = get_report(analytics,body)
-  ga_dict=print_response(response)
+    response = get_report(analytics, body)
+    ga_dict = print_response(response)
 
-  result=[]
-  for elmt in ga_dict:
-    print(elmt)
-    hour = datetime.now().strftime('%H')
-    # if int(hour)+1 > 8 :
-    message = send_msg_pg(elmt['ga:pageviews'])
-#      result.append(elmt)
-  print('inserting.....')
-  return message
+    result = []
+    for elmt in ga_dict:
+        print(elmt)
+        hour = datetime.now().strftime('%H')
+        # if int(hour)+1 > 8 :
+        message = send_msg_pg(viewid, elmt['ga:pageviews'])
+    #      result.append(elmt)
+    print('inserting.....')
+    return message
 
 
 if __name__ == '__main__':
-  creat_table()
-  main('123')
+    creat_table()
+    main('123')

+ 13 - 16
monitor/notifyall.py

@@ -11,40 +11,37 @@ from selenium.webdriver.common.by import By
 from selenium.webdriver.support import expected_conditions as EC
 import codecs
 import random
-from bs4 import BeautifulSoup
 import requests
 import dataset
 import time
-import rpyc
 import sys
 import docker
 import pymysql
 pymysql.install_as_MySQLdb()
-from monitor import GA_all
+from GA_all import main
 
 
-ga_info = {'班尼斯':['154114800',{
-        "Authorization": "Bearer " + "QCAM5upFjeBVp54PGqT4eMZSXPU0y4vYk1e1CoASa2P",
-        "Content-Type": "application/x-www-form-urlencoded"
-},'/Users/zooeytsai/Downloads/corded-velocity-301807-a3e3d5420aba.json'
-],'有夠讚':['208868237',{
+ga_info = {'有夠讚':['208868237',{
         "Authorization": "Bearer " + "Alur8iUgifYsh9CMiyUdlrTWK4Y9Xys8RZD52YkBMGl",
         "Content-Type": "application/x-www-form-urlencoded"
-},'/Users/zooeytsai/Downloads/corded-velocity-301807-a0d690cd59d9.json'],
-          '神助物流':['265874206',{
-        "Authorization": "Bearer " + "50Qu8O3XUhOhIOkz5WHy6mhftRVBe8CtL3DXS05UXRC",
-        "Content-Type": "application/x-www-form-urlencoded"},'/Users/zooeytsai/Downloads/corded-velocity-301807-42d10dfc21a4.json']}
+},'c:\gitlab\kw_tools\monitor\corded-velocity-301807-a0d690cd59d9.json'],
+        '新飛':['195216524',{"Authorization": "Bearer " + "ZfOUonutteqyFGfcx8jsHo9LoaTKL8oJ3WLwQamzN6c",
+        "Content-Type": "application/x-www-form-urlencoded"},'C:\gitlab\kw_tools\monitor\corded-velocity-301807-f43006dd42c8.json'],
+        '極簡乳清':['141324760',{"Authorization": "Bearer " + "DC7eiXkGoKzXcuCTuqCGmuNnDC15D3iv6E7tu5BAsUZ",
+        "Content-Type": "application/x-www-form-urlencoded"},'C:\gitlab\kw_tools\monitor\corded-velocity-301807-879b27e60620.json'],
+}
 
 def send_msg(viewid,header,key_file):
-    send_message = GA_all.main(viewid, key_file)
+    send_message = main(viewid, key_file)
     params = {"message": send_message['message']}
     print('通知結果',params)
     r = requests.post("https://notify-api.line.me/api/notify",headers=header, params=params)
 
 
-for k,v in ga_info.items():
-    send_msg(v[0],v[1],v[2])
-
+# for k,v in ga_info.items():
+# send_msg(v[0],v[1],v[2])
+send_msg('195216524',{"Authorization": "Bearer " + "DC7eiXkGoKzXcuCTuqCGmuNnDC15D3iv6E7tu5BAsUZ",
+        "Content-Type": "application/x-www-form-urlencoded"},"C:\/Users\/s1301\/Downloads\/corded-velocity-301807-f43006dd42c8.json")
 
 # schtasks /create /sc minute /mo 30 /sd 2022/05/050 /st 9:00 /et 23:00 /tn "linebat" /tr "C:\tmp\inn_line.bat"
 

+ 287 - 0
oak_u/etoday_use.py

@@ -0,0 +1,287 @@
+#import redis
+import time
+import traceback
+#import json
+from selenium import webdriver
+from selenium.webdriver.common.desired_capabilities import DesiredCapabilities
+import time
+import urllib
+import os
+from selenium.webdriver.support.ui import WebDriverWait
+from selenium.webdriver.common.by import By
+from selenium.webdriver.support import expected_conditions as EC
+import dataset
+from selenium.webdriver.common.keys import Keys
+import json
+import random
+import time
+import redis
+import sys
+import codecs
+import random
+import os
+import time
+import requests
+import datetime
+driver=None
+from fake_useragent import UserAgent
+
+ua = UserAgent()
+#proxy_enabled=True
+# proxy_enabled=False
+
+# # https://youtu.be/cR2M5Khgxvc
+
+# db = dataset.connect('mysql://choozmo:pAssw0rd@db.ptt.cx:3306/seo?charset=utf8mb4')
+# glog_table=db['general_log']
+
+def re_get_webdriver():
+    # global port
+    global driver
+    global portnum
+    # os.system('killall chrome')
+    result=[]
+    # if driver is not None:
+    #     print('closing....')
+    #     driver.quit()
+    #     print('quit....')
+    #     driver=None
+    # os.system()
+    options = webdriver.ChromeOptions()
+    options.add_argument("--user-agent=" +ua.random)    
+    options.add_argument("--no-sandbox")
+    options.add_argument("--headless")
+    options.add_argument("--incognito")
+    driver = webdriver.Remote(
+                command_executor='http://127.0.0.1:'+str(portnum)+'/wd/hub',
+            options=options)
+    return driver
+    # try:
+    #     options = webdriver.ChromeOptions()
+        
+    #     options.add_argument("--no-sandbox")
+    #     options.add_argument("--headless")
+    #     options.add_argument("--incognito")
+    #     # if proxy_enabled:
+    #     #     options.add_argument('--proxy-server=socks5://172.104.92.245:14900')
+
+    #     try:
+    #         driver = webdriver.Remote(
+    #             command_executor='http://127.0.0.1:'+str(portnum)+'/wd/hub',
+    #         options=options)
+    #     except:
+    #         traceback.print_exc()
+    #         return None
+    #     return driver
+    # except:
+    #     traceback.print_exc()
+    #     driver=None
+    #     return None
+    # return driver
+
+
+def run_once():
+    global count
+    global portnum
+    global bok
+    # global glog_table
+    # table=db['nda_log']
+    # print(jsobj)
+    # kw=jsobj['kw']
+
+    
+    # options = webdriver.ChromeOptions()
+    
+    # options.add_argument("--no-sandbox")
+    # options.add_argument("--headless")
+    # options.add_argument("--incognito")
+    # driver = webdriver.Remote(
+    #         command_executor='http://127.0.0.1:'+str(portnum)+'/wd/hub',
+    #     options=options)
+    
+        # if driver is not None:
+        #     break
+    
+    
+    ettoday_url_list = ['https://house.ettoday.net/news/1492047',
+        'https://house.ettoday.net/news/1492167',
+        'https://house.ettoday.net/news/1492288',
+        'https://house.ettoday.net/news/1492178',
+        'https://house.ettoday.net/news/1492229',
+        'https://house.ettoday.net/news/1492134',
+        'https://house.ettoday.net/news/1492240',
+        'https://house.ettoday.net/news/1492161',
+        'https://house.ettoday.net/news/1492168',
+        'https://house.ettoday.net/news/1492217']
+    
+    # try:        
+    for i in ettoday_url_list:
+        try:
+            driver=re_get_webdriver()
+        except:
+            portnum=random.randint(4555,4666)
+            print(portnum)
+            os.system('docker container stop p8816')
+            time.sleep(5)
+            os.system('docker container rm p8816')
+            time.sleep(5)
+            os.system('docker run -d -p '+str(portnum)+':4444 --name p8816 --dns 168.95.1.1 selenium/standalone-chrome:101.0')
+            bok += 1
+            count=0
+            time.sleep(5)
+            driver=re_get_webdriver()
+        time.sleep(3)
+        try:
+            driver.get(i)
+            time.sleep(3)
+            elmt_next = driver.find_element(By.XPATH, '//*[@id="house"]/div[3]/div[2]/div[6]/div/div/div[1]/article/div/div[3]/p[1]/a')
+
+            webdriver.ActionChains(driver).move_to_element(elmt_next).perform()
+            time.sleep(3)
+
+            webdriver.ActionChains(driver).move_to_element(elmt_next).click().perform()
+            print("cick!",i)
+            count+=1
+            print("count_time:",count,';borken_time:',bok)
+        # elmt = driver.find_element(By.XPATH, '//*[@id="yschsp"]')
+            time.sleep(random.randint(3,7))
+            driver.quit()
+        except:
+            driver.quit()
+            print("wrong",i,';borken_time:',bok)
+            time.sleep(5)
+    # except:
+    #     print('wrong for:',i)
+        # kw=jsobj['kw']
+        # if jsobj.get('domain') is None:
+        #     exclude=jsobj['exclude']
+        #     domain=None
+        # else:
+        #     domain=jsobj['domain']
+        #     exclude=None
+#         driver.get('https://www.google.com?num=100')
+#         time.sleep(17)
+#         while True:
+#             try:
+#                 print(driver.current_url)
+#                 break
+#             except:
+#                 traceback.print_exc()
+#                 driver=re_get_webdriver()
+#                 time.sleep(3)
+#                 driver.get('https://www.google.com?num=100')
+#                 time.sleep(3)
+
+#             time.sleep(3)
+
+#         elmt = driver.find_element(By.XPATH, "//input[@name='q']")
+#         time.sleep(1)
+#         elmt.send_keys(kw)
+#         elmt.send_keys(Keys.ENTER)
+#         time.sleep(6)
+
+#         elmts = driver.find_elements(By.XPATH, "//div[@class='yuRUbf']/a")
+
+#         numresults=len(elmts)
+# #        time.sleep(9999)
+#         print('搜尋結果數量',numresults)
+#         if numresults==0:
+#             print(driver.current_url)
+#             print(driver.title)
+#             sys.exit()
+
+#         idx=1
+#         found=False
+#         test_lst=[]
+#         for elmt in elmts:
+#             href=elmt.get_attribute('href')
+#             txt=elmt.text
+#             if len(txt)>10:
+#                 if domain is not None:
+#                     for d in domain:
+#                         if d in href:
+#                             print('found....')
+#                             print('clicked....')
+#                             print(href)
+#                             print(txt)
+#                             print("ranking", idx)
+#                             found=True
+
+#                             webdriver.ActionChains(driver).move_to_element(elmt).perform()
+# #                            elmt.click()
+#                             webdriver.ActionChains(driver).move_to_element(elmt).click().perform()
+#                             table.insert({'kw':kw,'ranking':idx,'title':txt,'url':href,'dt':datetime.datetime.now(),'result':numresults,'client':'64G'})
+#                             time.sleep(6)
+
+#                             return
+#                 else:
+#                     if exclude not in href:
+#                         test_lst.append(elmt)
+                    
+#             idx+=1
+#         if exclude is not None:
+#             print('exclude')
+#             elmt=random.choice(test_lst)
+#             print(elmt)
+
+#             webdriver.ActionChains(driver).move_to_element(elmt).perform()
+#             webdriver.ActionChains(driver).move_to_element(elmt).click().perform()
+#             time.sleep(5)
+
+#         if not found:
+#             table.insert({'ranking':-1,'kw':kw,'results':numresults,'url':'','title':'未收錄'})
+
+#     except:
+#         traceback.print_exc()
+
+#         print('exception')
+        # traceback.print_exc()
+        
+
+
+# time.sleep(5)
+
+# r=random.randint(0,27)
+# r=26
+
+# cursor=db.query('select json from seo_jobs where cust="KNIGHT" and plan="形象SEO" order by rand() limit 1')
+# for c in cursor:
+#     js=json.loads(c['json'])
+#     prefix=js['prefix']
+#     postfix=js['postfix']
+#     domain=js['domain'][0]
+#     positive=js['positive']
+#     rnd=js['rnd']
+
+
+portnum=random.randint(4555,4666)
+print(portnum)
+os.system('docker container stop p8816')
+time.sleep(5)
+os.system('docker container rm p8816')
+time.sleep(5)
+os.system('docker run -d -p '+str(portnum)+':4444 --name p8816 --dns 168.95.1.1 selenium/standalone-chrome:101.0')
+bok = 0
+count=0
+time.sleep(5)
+while True:
+    # run_once()
+    # time.sleep(10)
+
+    run_once()
+
+
+
+
+# kw=random.choice(positive)
+# kw2=random.choice(rnd)
+
+# count=0
+# while True:
+#     try:
+#         run_once({'domain':domain,'kw':prefix+" "+kw+" "+kw2})
+#         count+=1
+#     except:
+#         continue
+#     print('中場休息 次數',count)
+#     time.sleep(random.randint(120,150))

+ 1 - 0
oak_u/readme.md

@@ -0,0 +1 @@
+記得要修改chrome版本和docker名稱才能多個使用,另外雖然我有設定重啟,但時間長了還是會發生錯誤,並須重啟(暫時還不知道原因)

+ 377 - 0
oak_u/yahoo_use.py

@@ -0,0 +1,377 @@
+import time
+from datetime import datetime
+import json
+from selenium import webdriver
+from selenium.webdriver.common.desired_capabilities import DesiredCapabilities
+import time
+import os
+import urllib.parse
+from selenium.webdriver.support.ui import WebDriverWait
+from selenium.webdriver.common.by import By
+from selenium.webdriver.support import expected_conditions as EC
+import codecs
+import random
+import requests
+import dataset
+import traceback
+import sys
+from selenium.webdriver.common.keys import Keys
+import timeit
+import socket
+
+import random 
+import re
+
+
+# import requests
+
+
+
+from fake_useragent import UserAgent
+
+ua = UserAgent()
+def re_get_webdriver():
+    # global port
+    global driver
+    global portnum
+    # os.system('killall chrome')
+    result=[]
+    # if driver is not None:
+    #     print('closing....')
+    #     driver.quit()
+    #     print('quit....')
+    #     driver=None
+    # os.system()
+    options = webdriver.ChromeOptions()
+    options.add_argument("--user-agent=" +ua.random)    
+    options.add_argument("--no-sandbox")
+    options.add_argument("--headless")
+    options.add_argument("--incognito")
+    driver = webdriver.Remote(
+                command_executor='http://127.0.0.1:'+str(portnum)+'/wd/hub',
+            options=options)
+    return driver
+# headers = {'user-agent': ua.chrome}
+# r = requests.get('https://house.ettoday.net/news/1492047', headers=headers)
+# print(r.text)
+
+
+# options.binary_location = ('C:\\Program Files\\Google\\Chrome\\Application\\chrome.exe')
+# driverPath = './chromedriver.exe'
+
+# driver = webdriver.Firefox()
+# driver.get('https://google.com')
+
+# ettoday_url_list = ['https://house.ettoday.net/news/1492047',
+# 'https://house.ettoday.net/news/1492167',
+# 'https://house.ettoday.net/news/1492288',
+# 'https://house.ettoday.net/news/1492178',
+# 'https://house.ettoday.net/news/1492229',
+# 'https://house.ettoday.net/news/1492134',
+# 'https://house.ettoday.net/news/1492240',
+# 'https://house.ettoday.net/news/1492161',
+# 'https://house.ettoday.net/news/1492168',
+# 'https://house.ettoday.net/news/1492217']
+# for i in ettoday_url_list:
+    
+#     driver.get(i)
+#     time.sleep(3)
+#     elmt_next = driver.find_element(By.XPATH, '//*[@id="house"]/div[3]/div[2]/div[6]/div/div/div[1]/article/div/div[3]/p[1]/a')
+
+#     webdriver.ActionChains(driver).move_to_element(elmt_next).perform()
+#     webdriver.ActionChains(driver).move_to_element(elmt_next).click().perform()
+#     print("cick:",i)
+# # elmt = driver.find_element(By.XPATH, '//*[@id="yschsp"]')
+#     time.sleep(random.randint(3,7))
+#     driver.quit()
+
+# query='幸福空間'
+# elmt.send_keys(query)
+# elmt.send_keys(Keys.ENTER)
+# time.sleep(1)
+
+# time.sleep(1)
+
+def run_once():
+    global count
+    global bok
+    global portnum
+    yahoo_url_list = [
+    'https://house.yahoo.com.tw/%E9%9B%8D%E5%AE%B9%E9%9B%85%E7%B7%BB-%E5%84%AA%E9%9B%85%E5%81%87%E6%9C%9F-%E6%96%B0%E5%8F%A4%E5%85%B8-31%E5%9D%AA-020000499.html',
+    'https://house.yahoo.com.tw/%E6%96%B0%E7%94%9F%E9%AD%85%E5%8A%9B-%E8%AD%9C%E5%AF%AB%E5%B9%B8%E7%A6%8F%E5%9C%93%E8%88%9E%E6%9B%B2-%E5%8C%97%E6%AD%90%E9%A2%A8-35%E5%9D%AA-020000759.html',
+    'https://house.yahoo.com.tw/20%E5%B9%B4%E8%80%81%E5%AE%85%E9%87%8D%E7%94%9F-%E7%BE%8E%E5%BC%8F%E4%BD%8E%E5%A5%A2%E6%9C%89%E5%AE%B6%E7%9A%84%E6%BA%AB%E5%BA%A6-106%E5%9D%AA-020000087.html',
+    'https://house.yahoo.com.tw/sheer-%E7%B4%94%E7%B2%B9-%E7%8F%BE%E4%BB%A3%E9%A2%A8-25%E5%9D%AA-020000325.html',
+    'https://house.yahoo.com.tw/%E8%AE%8A%E5%BD%A2%E8%88%87%E7%B5%84%E5%90%88-%E8%A4%87%E5%90%88%E5%BC%8F%E7%9A%84%E7%A9%BA%E9%96%93%E8%A8%AD%E8%A8%88-%E4%B8%AD-020000869.html',
+    'https://house.yahoo.com.tw/%E8%A7%A3%E6%94%BE%E6%8B%98%E7%A6%81%E5%BF%83%E9%9D%88-%E8%B6%85%E8%84%AB%E7%8B%82%E6%83%B3%E9%80%8F%E5%A4%A9%E5%8E%9D-020000093.html',
+    'https://house.yahoo.com.tw/%E8%A6%AA%E5%AD%90%E6%96%99%E7%90%86%E7%9B%B4%E6%92%AD%E4%B8%BB%E7%9A%84%E5%AE%B6-%E5%BE%AE%E7%BE%8E%E5%BC%8F%E8%A8%AD%E8%A8%88-50%E5%9D%AA-020000607.html',
+    'https://house.yahoo.com.tw/%E5%82%B3%E9%81%94%E6%B7%B1%E8%89%B2%E6%BA%AB%E5%BA%A6-%E8%8B%B1%E5%80%AB%E7%B4%B3%E5%A3%AB%E8%B2%B4%E6%97%8F%E9%A2%A8-%E7%8F%BE%E4%BB%A3%E5%A5%A2%E8%8F%AF%E9%A2%A8-020000334.html',
+    'https://house.yahoo.com.tw/%E7%8E%A9%E5%91%B3%E7%B3%BB%E7%B5%B1%E6%9D%BF-%E5%BF%AB%E9%80%9F%E6%88%90%E5%AE%B6%E7%B0%A1%E7%B4%84%E7%8F%BE%E4%BB%A3%E9%A2%A8-35%E5%9D%AA-020000199.html',
+    'https://house.yahoo.com.tw/%E4%BB%A5%E5%9C%93%E5%BD%A2%E7%AC%A6%E7%A2%BC-%E5%BD%A2%E5%A1%91%E6%81%A2%E5%BC%98%E5%A5%A2%E7%BE%8E%E8%87%BB%E9%82%B8-%E5%A5%A2%E8%8F%AF%E9%A2%A8-42%E5%9D%AA-020000780.html']
+    
+    
+    for i in yahoo_url_list:
+        try:
+            try:
+                driver = re_get_webdriver()
+            except:
+                print('driver_bok')
+                portnum=random.randint(4555,4666)
+                print(portnum)
+                os.system('docker container stop p8809')
+                time.sleep(5)
+                os.system('docker container rm p8809')
+                time.sleep(5)
+                os.system('docker run -d -p '+str(portnum)+':4444 --name p8809 --dns 168.95.1.1 selenium/standalone-chrome:106.0')
+                count=0
+                bok+=1
+                time.sleep(5)
+                driver = re_get_webdriver()
+
+            driver.get(i)
+            time.sleep(5)
+            elmt_next = driver.find_element(By.XPATH, '//*[@id="maincontainer"]/main/div/div[2]/div[1]/div[1]/div[1]/div[1]/div/div/div[1]/a')
+
+            webdriver.ActionChains(driver).move_to_element(elmt_next).perform()
+            webdriver.ActionChains(driver).move_to_element(elmt_next).click().perform()
+            print("cick!")
+            count+=1
+            print('click_all_time:',count,';broken_time:',bok)
+        # elmt = driver.find_element(By.XPATH, '//*[@id="yschsp"]')
+            time.sleep(random.randint(3,7))
+            driver.quit()
+        except:
+            driver.quit()
+            print(i,'error',';broken_time:',bok)
+            time.sleep(10)
+portnum=random.randint(4555,4666)
+print(portnum)
+os.system('docker container stop p8809')
+time.sleep(5)
+os.system('docker container rm p8809')
+time.sleep(5)
+os.system('docker run -d -p '+str(portnum)+':4444 --name p8809 --dns 168.95.1.1 selenium/standalone-chrome:106.0')
+count=0
+bok=0
+time.sleep(5)
+while True:
+    # run_once()
+    # time.sleep(10)
+
+    run_once()
+# elmts=driver.find_elements("xpath",'//*[@id="web"]/ol/li/div/div[1]/h3/a')
+# domain = 'hhh.com.tw'
+# idx=1
+# ranking=-1
+# domain_in_link = 0
+# print (len(elmts))
+# # driver.save_screenshot('c:/tmp/test.png')
+# n=0
+# for el in elmts:
+#     n+=1
+#     href=el.get_attribute('href')
+#     txt=el.text
+#     # print(txt)
+#     if len(txt)>10:
+#         if domain in href:
+#             domain_in_link += 1
+#             print('clicked....')
+#             print('href:',href)
+#             print('txt:',txt)
+# elmt_next = driver.find_element(By.XPATH, '//*[@id="left"]/div/ol/li[1]/div/div/a')
+
+# webdriver.ActionChains(driver).move_to_element(elmt_next).perform()
+# webdriver.ActionChains(driver).move_to_element(elmt_next).click().perform()
+# time.sleep(2)
+# elmts=driver.find_elements("xpath",'//*[@id="web"]/ol/li/div/div[1]/h3/a')
+# domain = 'hhh.com.tw'
+# idx=1
+# ranking=-1
+# domain_in_link = 0
+# print (len(elmts))
+# # driver.save_screenshot('c:/tmp/test.png')
+# n=0
+# for el in elmts:
+#     n+=1
+#     href=el.get_attribute('href')
+#     txt=el.text
+#     # print(txt)
+#     if len(txt)>10:
+#         if domain in href:
+#             domain_in_link += 1
+#             print('clicked....')
+#             print('href:',href)
+#             print('txt:',txt)
+# elmt_next = driver.find_element(By.XPATH, '//*[@id="left"]/div/ol/li[1]/div/div/a[2]')
+
+# webdriver.ActionChains(driver).move_to_element(elmt_next).perform()
+# webdriver.ActionChains(driver).move_to_element(elmt_next).click().perform()
+# time.sleep(5)
+# for i in range(20):
+#     try:
+#         elmt_next = driver.find_element(By.XPATH, '//*[@id="left"]/div/ol/li[1]/div/div/a[2]')
+
+#         webdriver.ActionChains(driver).move_to_element(elmt_next).perform()
+#         webdriver.ActionChains(driver).move_to_element(elmt_next).click().perform()
+#         time.sleep(5)
+#     except:
+#         time.sleep(200)
+#                 webdriver.ActionChains(driver).move_to_element(el).click().perform()
+# add_tabs = [7,9,11,13,15,7,9,11,13,15,7,9,11,13,15,7,9,11,13,15]
+
+# db = dataset.connect('mysql://choozmo:pAssw0rd@db.ptt.cx:3306/seo?charset=utf8mb4')
+# driver=None
+# headers = {
+#         "Authorization": "Bearer " + "t35vhZtWNgvDNWHc3DJh0OKll3mcB9GvC8K2EAkBug2",
+#         "Content-Type": "application/x-www-form-urlencoded"
+# }
+
+# sleepoffset = 0
+
+# def send_msg(kw):
+#     params = {"message": "處理關鍵字: "+kw}  
+#     r = requests.post("https://notify-api.line.me/api/notify",headers=headers, params=params)
+
+# def empty_query(q):
+#     global driver
+#     googleurl='https://www.google.com/search?q='+urllib.parse.quote(q)
+#     driver.get(googleurl)
+#     time.sleep(3)
+
+# def process_query(domain, target_domain, brands, query):
+#     print(query)
+#     sleepoffset = 0
+#     global driver
+#     if query == "艾立思" and "index" in target_domain:
+#         driver.get('https://www.google.com/search?num=100&q=艾立思&rlz=1C1ONGR_zh-TWTW997TW997&ei=zjdUY_DBG9Lm-Abpgq84&start=0&sa=N&filter=0&ved=2ahUKEwjw4KeEvfT6AhVSM94KHWnBCwcQ8tMDegQIARAQ&cshid=1666463754367857&biw=1368&bih=761&dpr=2')
+#         time.sleep(4)
+#     else:
+#         driver.get('https://www.google.com?num=100')
+#         time.sleep(3)
+#         print(driver.current_url)
+
+#         # elmts=driver.find_elements_by_xpath("//div[@class='yuRUbf']/a")
+#         # ABOVE METHOD IS DEPRECATED STARTING SELENIUM 4.3.0, USE THIS
+#         #
+#         elmt = driver.find_element(By.XPATH, "//input[@name='q']")
+#         time.sleep(1)
+
+#         elmt.send_keys(query)
+#         elmt.send_keys(Keys.ENTER)
+
+#     idx=1
+#     ranking=-1
+#     domain_in_link = 0
+
+#     googleurl = driver.current_url
+#     print(driver.current_url)
+
+#     if "sorry" in googleurl:
+#         return 444
+
+#     elmts=driver.find_elements("xpath","//div[@class='yuRUbf']/a")
+
+
+#     print (len(elmts))
+#     # driver.save_screenshot('c:/tmp/test.png')
+#     n=0
+#     for el in elmts:
+#         n+=1
+#         href=el.get_attribute('href')
+#         txt=el.text
+#         if len(txt)>10:
+#             if domain in href:
+#                 domain_in_link += 1
+#                 print('clicked....')
+#                 print(href)
+#                 print(txt)
+
+#                 if query == "艾立思" and "index" in target_domain and href != "https://hhh.com.tw/brand-index.php?brand_id=211":
+#                     print("wrong site")
+#                     continue
+                
+#                 webdriver.ActionChains(driver).move_to_element(el).perform()
+#                 webdriver.ActionChains(driver).move_to_element(el).click().perform()
+#                 print("Rank: " + str(n))
+#                 time.sleep(15)
+
+#                 ''' unused
+#                 new_windows_count = add_tabs[random.randint(0,19)]
+#                 print(str(new_windows_count) + " new tabs")
+#                 for i in range (0,new_windows_count):
+#                     print("Tab " + str(i+1))
+#                     #original_window = driver.current_window_handle
+#                     #driver.switch_to.new_window('window')
+#                     #driver.get(href)
+#                     sleepoffset += 12
+#                     driver.execute_script('window.open("'+href+'","_blank");')
+#                     driver.execute_script("window.scrollTo(0, 600)")
+#                     time.sleep(15)
+#                     #driver.close()
+#                     #driver.switch_to.window(original_window)
+                
+#                 if domain in target_domain:
+#                     print("Target link found")
+#                     time_stamp = datetime.fromtimestamp(time.time())
+#                     time_stamp = time_stamp.strftime("%Y-%m-%d %H:%M:%S")
+#                     db['click_results'].insert({"time_stamp": time_stamp, "brand": brands[domain], "domain": domain, "query": query, "url": href, "content": txt, "extra_windows": '0'})
+#                 '''
+#                 break
+
+#     '''if domain in target_domain:
+#         print("Target domain found")
+#         time_stamp = datetime.fromtimestamp(time.time())
+#         time_stamp = time_stamp.strftime("%Y-%m-%d %H:%M:%S")
+#         db['query_results'].insert({"time_stamp": time_stamp, "brand": brands[domain], "domain": domain, "query": query, "googleurl": googleurl, "element_count": len(elmts), "domain_in_link_count": domain_in_link})
+#     '''        
+
+#     print(domain_in_link)
+#     return 200
+    
+
+# def run_once(domain, target_domain, brands, query):
+#     global driver
+#     result=[]
+#     options = webdriver.ChromeOptions()
+#     options.add_argument('--headless')
+# #    options.add_argument("--user-agent=" +user_agent)
+#     options.add_argument("--incognito")
+#     options.add_argument('--no-sandbox')
+#     options.add_argument('--disable-dev-shm-usage')
+
+#     driver = webdriver.Chrome(
+#     options=options)
+
+#     driver.delete_all_cookies()
+#     driver.set_window_size(1400,1000)
+
+#     statuscode = process_query(domain, target_domain, brands, query)
+#     driver.quit()
+
+#     return statuscode
+
+# #execution starts here
+
+# def execute(domain, target_domain, brands, query_list):
+#     print("Ctrl+C or Ctrl+Z to stop.")
+#     statuscode = 0
+#     st = timeit.default_timer()
+#     try:
+#         statuscode = run_once(domain, target_domain, brands, random.choice(query_list))
+#     except:
+#         traceback.print_exc()
+#     timetaken = timeit.default_timer()-st
+#     print("Time taken: " + str(timetaken))
+    
+#     print("Process returned with " + str(statuscode))
+#     if statuscode == 444:
+#         print("You have been caught!!!")
+        
+#         #notify("Clickbot " + brands[domain] + " has been caught by Google and will terminate. IP: ")
+
+#     extrasleep = 0
+#     if(timetaken < 50):
+#         extrasleep = 50 - timetaken
+#     print("Ctrl+C or Ctrl+Z to stop now.")
+#     print("You have " + str(10 + extrasleep) + " seconds.")
+#     time.sleep(10 + extrasleep)
+#     return statuscode