zooey 2 years ago
parent
commit
f6aefa6ccb

+ 0 - 12
INNNews/general_clickbot_local.py

@@ -11,7 +11,6 @@ from selenium.webdriver.chrome.service import Service
 from selenium.webdriver.support import expected_conditions as EC
 from selenium.webdriver.support import expected_conditions as EC
 import codecs
 import codecs
 import random
 import random
-import requests
 import datetime
 import datetime
 import dataset
 import dataset
 import pymysql
 import pymysql
@@ -115,15 +114,4 @@ if __name__ == '__main__':
     fire.Fire(JParams)
     fire.Fire(JParams)
 
 
 
 
-cursor=db.query('select json from seo_jobs where cust="KNIGHT" and plan="形象SEO" order by rand() limit 1')
-for c in cursor:
-    js=json.loads(c['json'])
-    prefix=js['prefix']
-    postfix=js['postfix']
-    domain=js['domain'][0]
-    positive=js['positive']
-    rnd=js['rnd']
 
 
-kw1=random.choice(positive)
-kw2=random.choice(rnd)
-kw=prefix+" "+kw1+" "+kw2

+ 10 - 20
INNNews/run_sheet_local.py

@@ -1,16 +1,9 @@
 import sys
 import sys
 import codecs
 import codecs
 import traceback
 import traceback
-import requests
 import re
 import re
 import pandas as pd
 import pandas as pd
 import random
 import random
-import urllib
-import json
-import gspread
-import datetime
-from gspread_pandas import Spread, Client
-from oauth2client.service_account import ServiceAccountCredentials
 import os
 import os
 import redis
 import redis
 import time
 import time
@@ -31,7 +24,7 @@ def run_once():
     kw = obj
     kw = obj
     domain = 'jiatetea.tw'
     domain = 'jiatetea.tw'
     print(kw, domain)
     print(kw, domain)
-    os.chdir('/Users/zooeytsai/kw_tools/INNNews')
+    os.chdir('C:\/Users\/s1301\/PycharmProjects\/kw_tools\/INNNews')
     intval = os.system(f'python3 general_clickbot_local.py get --kw="{kw}" --domain="{domain}"')
     intval = os.system(f'python3 general_clickbot_local.py get --kw="{kw}" --domain="{domain}"')
     
     
     print('執行完成genetal_clickbot')
     print('執行完成genetal_clickbot')
@@ -46,19 +39,16 @@ def run_specific():
     db = dataset.connect('mysql://choozmo:pAssw0rd@db.ptt.cx:3306/seo?charset=utf8mb4')
     db = dataset.connect('mysql://choozmo:pAssw0rd@db.ptt.cx:3306/seo?charset=utf8mb4')
     # db_local = dataset.connect('mysql://root:jondae350@localhost/ig_tags')
     # db_local = dataset.connect('mysql://root:jondae350@localhost/ig_tags')
     lst = []
     lst = []
-    
-    # cursor = db.query('select term, domain from ig_tags.select_kw where client="班尼斯口碑"')
-    # for c in cursor:
-    #     lst.append([c['term'], c['domain']])
-    
-    # obj = random.choice(lst)
-    # kw = obj[0]
-    kw_list = ['倉儲 管理']
-    kw = random.choice(kw_list)
-    # domain = obj[1]
-    domain = 'gs-rack.com'
+
+    cursor = db.query('select term, domain from seo.selected_kw where client="極簡乳清"')
+    for c in cursor:
+        lst.append([c['term'], c['domain']])
+
+    obj = random.choice(lst)
+    kw = obj[0]
+    domain = obj[1]
     print(kw, domain)
     print(kw, domain)
-    os.chdir('/Users/zooeytsai/kw_tools/INNNews')
+    # os.chdir('/Users/zooeytsai/kw_tools/INNNews')
     intval = os.system(f'python3 general_clickbot_local.py get --kw="{kw}" --domain="{domain}"')
     intval = os.system(f'python3 general_clickbot_local.py get --kw="{kw}" --domain="{domain}"')
     print('執行完成genetal_clickbot')
     print('執行完成genetal_clickbot')
     
     

+ 10 - 9
SEO/clickbot_100.py

@@ -45,15 +45,16 @@ def process_one():
     db = dataset.connect('mysql://choozmo:pAssw0rd@db.ptt.cx:3306/seo?charset=utf8mb4')
     db = dataset.connect('mysql://choozmo:pAssw0rd@db.ptt.cx:3306/seo?charset=utf8mb4')
     lst=[]
     lst=[]
     table=db['google_rank']
     table=db['google_rank']
-    # cursor = db.query('select term,domain from seo.selected_kw where client="護全"')
+    cursor = db.query('select term,domain from seo.selected_kw where client="misa"')
     # cursor = db.query('select term,url from seo.sns_kw where client="loginheart"')
     # cursor = db.query('select term,url from seo.sns_kw where client="loginheart"')
     # cursor = db.query('select term from seo.selected_kw where client="鑫富"')
     # cursor = db.query('select term from seo.selected_kw where client="鑫富"')
-    cursor = db.query('select term,domain from seo.selected_kw where id between 2146 and 2155')
+    # cursor = db.query('select term,domain from seo.select_kw where id between 1148 and 1155')
+    # cursor = db.query('select term,domain from seo.select_kw where client="有夠讚"')
     # cursor=db.query('select term from selected_kw and term not in (SELECT distinct(keyword) FROM ig_tags.save_result where url like "%beastparadise.net%" and datediff(now(),dt)=0)')
     # cursor=db.query('select term from selected_kw and term not in (SELECT distinct(keyword) FROM ig_tags.save_result where url like "%beastparadise.net%" and datediff(now(),dt)=0)')
     for c in cursor:
     for c in cursor:
         lst.append([c['term'],c['domain']])
         lst.append([c['term'],c['domain']])
 
 
-    # domain = 'wandatw.com'
+    # domain = 'yogoclean.com'
     for i in lst:
     for i in lst:
         print(i)
         print(i)
         driver=restart_browser()
         driver=restart_browser()
@@ -62,7 +63,7 @@ def process_one():
         driver.get(googleurl)
         driver.get(googleurl)
         time.sleep(60)
         time.sleep(60)
         print(driver.current_url)
         print(driver.current_url)
-        driver.save_screenshot('C:\/Users\/s1301\/Pictures\/Saved Pictures\/護全\/'+i[0]+'.png')
+        driver.save_screenshot('C:\/Users\/s1301\/Pictures\/Saved Pictures\/misa\/'+i[0]+'.png')
         df=pd.DataFrame()
         df=pd.DataFrame()
         elmts=driver.find_elements(By.XPATH,"//div[@class='yuRUbf']/a")
         elmts=driver.find_elements(By.XPATH,"//div[@class='yuRUbf']/a")
         print('結果數量',len(elmts))
         print('結果數量',len(elmts))
@@ -76,10 +77,10 @@ def process_one():
                     table.insert({'title':elmt.text,'url':href,'keyword':i[0],'dt':datetime.datetime.now(),'ranking':cnt})
                     table.insert({'title':elmt.text,'url':href,'keyword':i[0],'dt':datetime.datetime.now(),'ranking':cnt})
                     print(href)
                     print(href)
                     print(elmt.text)
                     print(elmt.text)
-                # datadict['搜尋詞'].append(term)
-                # datadict['結果標題'].append(elmt.text)
-                # datadict['結果網址'].append(href)
-                # datadict['結果名次'].append(str(cnt))
+                datadict['搜尋詞'].append(i[0])
+                datadict['結果標題'].append(elmt.text)
+                datadict['結果網址'].append(href)
+                datadict['結果名次'].append(str(cnt))
 
 
                 cnt+=1
                 cnt+=1
             except:
             except:
@@ -92,7 +93,7 @@ def process_one():
         df['結果網址']=datadict['結果網址']
         df['結果網址']=datadict['結果網址']
         df['結果名次']=datadict['結果名次']
         df['結果名次']=datadict['結果名次']
 
 
-        df.to_excel('C:\/Users\/s1301\/Pictures\/Saved Pictures\/護全\/'+i[0]+".xls")
+        df.to_excel('C:\/Users\/s1301\/Pictures\/Saved Pictures\/misa\/'+i[0]+".xls")
 
 
         driver.quit()
         driver.quit()
         print('等待')
         print('等待')

+ 4 - 2
SEO/clickbot_100_one.py

@@ -41,8 +41,10 @@ def restart_browser():
     return driver
     return driver
 
 
 def process_one():
 def process_one():
-    lst=['龔國權']
-    date='1223'
+    # lst=['信義房屋','信義 房屋','信義房仲','信義 房仲']
+    lst=['台北 完美電波']
+    # lst = ['龔國權']
+    date=''
     for term in lst:
     for term in lst:
         driver=restart_browser()
         driver=restart_browser()
         escaped_search_term=urllib.parse.quote(term)
         escaped_search_term=urllib.parse.quote(term)

+ 21 - 17
SEO/csv_to_sql.py

@@ -7,37 +7,41 @@ pymysql.install_as_MySQLdb()
 
 
 
 
 db = dataset.connect('mysql://choozmo:pAssw0rd@db.ptt.cx:3306/seo?charset=utf8mb4')
 db = dataset.connect('mysql://choozmo:pAssw0rd@db.ptt.cx:3306/seo?charset=utf8mb4')
-table=db['seo_jobs']
-# table=db['selected_kw']
+# table=db['seo_jobs']
+table=db['selected_kw']
 # table=db['sns_kw']
 # table=db['sns_kw']
-client='信義房屋'
-domain='www.hujiai-international.com'
+# table=db['select_kw']
+client='美麗馨'
+domain='lisinbeauty.com'
 # for index,row in df.iterrows():
 # for index,row in df.iterrows():
 # with open("C:\/Users\/s1301\/Documents\/新飛國際遊學SEO - 關鍵字12.08.csv") as f:
 # with open("C:\/Users\/s1301\/Documents\/新飛國際遊學SEO - 關鍵字12.08.csv") as f:
 #     data_all = f.readlines()
 #     data_all = f.readlines()
 # print(data_all)
 # print(data_all)
 f = open("C:\/Users\/s1301\/Documents\/456.csv",encoding='utf-8')
 f = open("C:\/Users\/s1301\/Documents\/456.csv",encoding='utf-8')
 # df = pd.read_csv(f,header=None, names=['kw', 'url'])
 # df = pd.read_csv(f,header=None, names=['kw', 'url'])
-# df = pd.read_csv(f,header=None, names=['kw'])
-df = pd.read_csv(f,header=None, names=['prefix', 'positive','rnd','domain'])
+df = pd.read_csv(f,header=None, names=['kw'])
+# df = pd.read_csv(f,header=None, names=['prefix','id', 'positive','domain','rnd'])
 df=df.fillna('')
 df=df.fillna('')
 # print(df)
 # print(df)
 # url='fflc.tw'
 # url='fflc.tw'
 lst=[]
 lst=[]
 for index,row in df.iterrows():
 for index,row in df.iterrows():
-    # print(row['kw'],url)
     # print(row['kw'])
     # print(row['kw'])
     # lst.append(row['kw'])
     # lst.append(row['kw'])
-    prefix='"'+row['prefix']+'"'
-    positive='"'+row['positive']+'"'
-    domain='"'+row['domain']+'"'
-    rnd='"'+str(row['rnd'])+'"'
-    postfix='""'
-    data = f'"prefix":{prefix},"domain":[{domain}],"postfix":{postfix},"positive":[{positive}],"rnd":[{rnd}]'
-    json='{'+data+'}'
-    print(json)
-    table.insert({'cust':client,'plan':'形象SEO','json':json})
-    # table.insert({'term':row['kw'],'client':client,'domain':domain})
+    # prefix='"'+row['prefix']+'"'
+    # # positive='"'+row['positive']+'"'
+    # positive=row['positive']
+    # domain='"'+row['domain']+'"'
+    # rnd='"'+str(row['rnd'])+'"'
+    # postfix='""'
+    # id=row['id']
+    # data = f'"id":{id},"prefix":{prefix},"domain":[{domain}],"postfix":{postfix},"positive":[{positive}],"rnd":[{rnd}]'
+    # json='{'+data+'}'
+    # print(json)
+    #
+    # table.insert({'cust':client,'plan':'形象SEO','json':json})
+    table.insert({'term':row['kw'],'client':client,'domain':domain})
+    # table.insert({'term': row['kw'], 'client': client, 'url': row['url']})
 print(lst)
 print(lst)
 {"prefix": "護佳", "postfix": "", "domain": ["www.hujiai-international.com"], "positive": ["細胞食物","紅蔘"], "rnd": [""]}
 {"prefix": "護佳", "postfix": "", "domain": ["www.hujiai-international.com"], "positive": ["細胞食物","紅蔘"], "rnd": [""]}
 
 

+ 3 - 3
SEO/month_kw_rank.py

@@ -6,9 +6,9 @@ def day_col(row):
     return result
     return result
 
 
 
 
-custom_name = ['清原']
+custom_name = ['瑞福']
 for name in custom_name:
 for name in custom_name:
-    df = pd.read_csv("C:\\Users\\s1301\\Documents\\11月清原排名前10_1.csv")
+    df = pd.read_csv(f"C:\\Users\\s1301\\Documents\\12月{name}排名前10.csv",encoding='utf-8')
     df['dt2'] = df.apply(day_col, axis=1)
     df['dt2'] = df.apply(day_col, axis=1)
     df = df.drop_duplicates(subset=['dt2','kw'])
     df = df.drop_duplicates(subset=['dt2','kw'])
     df_kw_rank = df[['dt2','kw','ranking']].sort_values(by='dt2')
     df_kw_rank = df[['dt2','kw','ranking']].sort_values(by='dt2')
@@ -34,4 +34,4 @@ for name in custom_name:
     # df_result.insert(0,'日期 ',df_kw_rank['日期'])
     # df_result.insert(0,'日期 ',df_kw_rank['日期'])
     # df_result.insert(1,'關鍵字 ',df_kw_rank['關鍵字'])
     # df_result.insert(1,'關鍵字 ',df_kw_rank['關鍵字'])
     # df_result.insert(2,'名次 ',df_kw_rank['名次'])
     # df_result.insert(2,'名次 ',df_kw_rank['名次'])
-    new.to_csv(f"C:\/Users\/s1301\/Documents\/{name}11月績效報表2.csv",index=False)
+    new.to_csv(f"C:\/Users\/s1301\/Documents\/{name}12月績效報表.csv",index=False)

+ 3 - 4
SEO/mysql_to_redis.py

@@ -8,14 +8,13 @@ import json
 db = dataset.connect('mysql://choozmo:pAssw0rd@db.ptt.cx:3306/seo?charset=utf8mb4')
 db = dataset.connect('mysql://choozmo:pAssw0rd@db.ptt.cx:3306/seo?charset=utf8mb4')
 table=db['selected_kw']
 table=db['selected_kw']
 lst = []
 lst = []
-cur = db.query('select term,url from seo.sns_kw where client="理茶"')
+cur = db.query('select term,domain from seo.selected_kw where client="plantA"')
 for c in cur:
 for c in cur:
-    lst.append([c['term'],c['url']])
-
+    lst.append([c['term'],c['domain']])
+db.close()
 
 
 print(lst)
 print(lst)
 # r = redis.Redis(host='db.ptt.cx', port=6379, db=1,password='choozmo9')
 # r = redis.Redis(host='db.ptt.cx', port=6379, db=1,password='choozmo9')
 # r.set(lst)
 # r.set(lst)
-['中 美 街 飲料', 'https://www.daisyyohoho.com/richa-drinks/'], ['中 美 街 飲料', 'https://www.walkerland.com.tw/article/view/340526'], ['中 美 街 飲料', 'https://www.facebook.com/Richa.drinks/'], ['中 美 街 飲料店', 'https://www.walkerland.com.tw/article/view/340526'], ['中 美 街 飲料 店', 'https://chiaoda.com/richa/'], ['勤美飲料店', 'https://playqueen888.com/tourall.php?twId=772&twlistId=1'], ['台中必喝飲料', 'https://playqueen888.com/tourall.php?twId=772&twlistId=1'], ['勤美 飲料 內用', 'https://www.walkerland.com.tw/article/view/340526'], ['精誠 路 飲料 店', 'https://chiaoda.com/richa/'], ['中 美 街 飲料 店推薦', 'https://zh-tw.facebook.com/Richa.drinks/'], ['台中中美街飲料', 'https://www.daisyyohoho.com/richa-drinks/'], ['中美街飲料推薦', 'https://www.daisyyohoho.com/richa-drinks/'], ['內用的手搖飲料店', 'https://taiwan17go.com/richa/'], ['台中限定飲料店', 'https://taiwan17go.com/richa/'], ['台中西區飲料推薦', 'https://chiaoda.com/richa/'], ['台中內用飲料店', 'https://chiaoda.com/richa/'], ['精誠路飲料', 'https://chiaoda.com/richa/'], ['台中中美街飲料', 'https://www.facebook.com/Richa.drinks/'], ['中美街 手搖飲', 'https://www.walkerland.com.tw/article/view/340526'], ['中美街 手搖飲', 'https://wandatw.com/richa/'], ['中美街手搖', 'https://www.daisyyohoho.com/richa-drinks/'], ['精誠路手搖', 'https://chiaoda.com/richa/']
 
 
 #####用記事本替換#######
 #####用記事本替換#######

+ 2 - 2
SEO/ranking_report.py

@@ -3,8 +3,8 @@ import dataset
 import pymysql
 import pymysql
 pymysql.install_as_MySQLdb()
 pymysql.install_as_MySQLdb()
 
 
-name='plantA'
-date = '1222'
+name='瑞福'
+date = '0110'
 
 
 
 
 def day():
 def day():

+ 19 - 17
SEO/ranking_world.py

@@ -88,7 +88,7 @@ def re_get_webdriver():
 
 
 
 
 def scrolling(driver, pgnum):
 def scrolling(driver, pgnum):
-    ub = driver.find_element_by_css_selector('body')
+    ub = driver.find_element("css selector",'body')
     for i in range(pgnum):
     for i in range(pgnum):
         ub.send_keys(Keys.PAGE_DOWN)
         ub.send_keys(Keys.PAGE_DOWN)
         if pgnum > 1:
         if pgnum > 1:
@@ -110,7 +110,7 @@ def run_once(jsobj):
     try:
     try:
         kw = jsobj['kw']
         kw = jsobj['kw']
         fname = jsobj['fname']
         fname = jsobj['fname']
-
+        url = jsobj['url']
         # if jsobj.get('domain') is None:
         # if jsobj.get('domain') is None:
         #     exclude = jsobj['exclude']
         #     exclude = jsobj['exclude']
         #     domain = None
         #     domain = None
@@ -130,24 +130,26 @@ def run_once(jsobj):
         })
         })
 
 
         # 芝加哥、邁阿密、紐約、華盛頓
         # 芝加哥、邁阿密、紐約、華盛頓
-        driver.execute_cdp_cmd("Emulation.setGeolocationOverride", Map_coordinates)
-        driver.get('https://www.google.com?num=100&lr=lang_en')
+        # driver.execute_cdp_cmd("Emulation.setGeolocationOverride", Map_coordinates)
+        # driver.get('https://www.google.com?num=100&lr=lang_en')
 
 
-        time.sleep(30)
-        driver.get('https://www.google.com?num=100&lr=lang_en')
-        print(driver.current_url)
-        elmt = driver.find_element(By.XPATH, "//input[@name='q']")
-        time.sleep(1)
-        elmt.send_keys(kw)
 
 
-        elmt.send_keys(Keys.ENTER)
-        time.sleep(6)
+        driver.get(url)
+        # print(driver.current_url)
+        # elmt = driver.find_element(By.XPATH, "//input[@name='q']")
+        # time.sleep(1)
+        # elmt.send_keys(kw)
+        #
+        # elmt.send_keys(Keys.ENTER)
+        time.sleep(3)
+        scrolling(driver, 10)
+        time.sleep(20)
 
 
         elmts = driver.find_elements(By.XPATH, "//div[@class='yuRUbf']/a")
         elmts = driver.find_elements(By.XPATH, "//div[@class='yuRUbf']/a")
         numresults = len(elmts)
         numresults = len(elmts)
         print('搜尋結果數量', numresults)
         print('搜尋結果數量', numresults)
         time.sleep(20)
         time.sleep(20)
-        driver.save_screenshot('C:\/Users\/s1301\/Pictures\/Saved Pictures\/angelo koo\/' +date +fname + ".png")
+        # driver.save_screenshot('C:\/Users\/s1301\/Pictures\/Saved Pictures\/angelo koo\/' +date +fname + ".png")
         if numresults == 0:
         if numresults == 0:
             send_msg('stop working...')
             send_msg('stop working...')
             sys.exit()
             sys.exit()
@@ -234,11 +236,11 @@ def run_once(jsobj):
     sys.exit()
     sys.exit()
 
 
 
 
-run_once({'kw':'angelo koo','fname':'chicago','date':'1222'})
+run_once({'kw':'angelo koo','fname':'chicago','date':'0112','url':'https://www.google.com/search?q=angelo+koo&hl=en&gl=us&num=100&uule=w+CAIQICIeQ2hpY2FnbyxJbGxpbm9pcyxVbml0ZWQgU3RhdGVz&gws_rd=cr'})
 
 
 ####手動截圖:須按右下角的設定選擇區域######
 ####手動截圖:須按右下角的設定選擇區域######
 ny="https://www.google.com/search?q=angelo+koo&hl=en&gl=us&num=100&uule=w+CAIQICIWTmV3IFlvcmssVW5pdGVkIFN0YXRlcw&gws_rd=cr"
 ny="https://www.google.com/search?q=angelo+koo&hl=en&gl=us&num=100&uule=w+CAIQICIWTmV3IFlvcmssVW5pdGVkIFN0YXRlcw&gws_rd=cr"
-wc="https://www.google.com/search?q=angelo+koo&hl=en&gl=us&uule=w+CAIQICIYV2FzaGluZ3RvbixVbml0ZWQgU3RhdGVz&gws_rd=cr"
-miami="https://www.google.com/search?q=angelo+koo&hl=en&gl=us&uule=w+CAIQICIbTWlhbWksRmxvcmlkYSxVbml0ZWQgU3RhdGVz&gws_rd=cr"
-chicago="https://www.google.com/search?q=angelo+koo&hl=en&gl=us&uule=w+CAIQICIeQ2hpY2FnbyxJbGxpbm9pcyxVbml0ZWQgU3RhdGVz&gws_rd=cr"
+wc="https://www.google.com/search?q=angelo%20koo&hl=en&gl=us&num=100&uule=w+CAIQICItV2FzaGluZ3RvbixEaXN0cmljdCBvZiBDb2x1bWJpYSxVbml0ZWQgU3RhdGVz&gws_rd=cr#gws_rd=cr&ip=1"
+miami="https://www.google.com/search?q=angelo+koo&hl=en&gl=us&num=100&uule=w+CAIQICIbTWlhbWksRmxvcmlkYSxVbml0ZWQgU3RhdGVz&gws_rd=cr"
+chicago="https://www.google.com/search?q=angelo+koo&hl=en&gl=us&num=100&uule=w+CAIQICIeQ2hpY2FnbyxJbGxpbm9pcyxVbml0ZWQgU3RhdGVz&gws_rd=cr"
 
 

+ 19 - 23
SEO/ranking_world_2.py

@@ -64,7 +64,7 @@ def re_get_webdriver():
         options = webdriver.ChromeOptions()
         options = webdriver.ChromeOptions()
         options.add_argument("--no-sandbox")
         options.add_argument("--no-sandbox")
         options.add_argument("--disable-dev-shm-usage")
         options.add_argument("--disable-dev-shm-usage")
-        # options.add_argument("--headless")
+        options.add_argument("--headless")
         print(user_agent)
         print(user_agent)
         options.add_experimental_option('prefs', {'intl.accept_languages': 'en,en_US'})
         options.add_experimental_option('prefs', {'intl.accept_languages': 'en,en_US'})
         options.add_argument("--incognito")
         options.add_argument("--incognito")
@@ -74,7 +74,7 @@ def re_get_webdriver():
                 driver = webdriver.Chrome(options=options)
                 driver = webdriver.Chrome(options=options)
 
 
             else:
             else:
-                driver = webdriver.Chrome(executable_path=r'C:\Users\Administrator\Downloads\chromedriver_108\chromedriver', options=options)
+                driver = webdriver.Chrome(executable_path=r'C:\Users\Administrator\Downloads\chromedriver_107\chromedriver', options=options)
         except:
         except:
             traceback.print_exc()
             traceback.print_exc()
             return
             return
@@ -153,7 +153,7 @@ def run_once(jsobj):
                               'dt': datetime.datetime.now(), 'client': 'i5'})
                               'dt': datetime.datetime.now(), 'client': 'i5'})
                 time.sleep(6)
                 time.sleep(6)
                 db.close()
                 db.close()
-
+                break
             idx += 1
             idx += 1
 
 
 
 
@@ -172,11 +172,7 @@ def run_once(jsobj):
         #     table.insert(
         #     table.insert(
         #         {'ranking': -1, 'kw': kw, 'results': numresults, 'url': '', 'title': '未收錄', 'descrption': desc})
         #         {'ranking': -1, 'kw': kw, 'results': numresults, 'url': '', 'title': '未收錄', 'descrption': desc})
 
 
-        df['搜尋詞'] = datadict['搜尋詞']
-        df['結果標題'] = datadict['結果標題']
-        df['結果網址'] = datadict['結果網址']
-        df['結果名次'] = datadict['結果名次']
-        df['結果說明'] = datadict['結果說明']
+
 
 
         # df.to_excel('C:\/Users\/s1301\/Pictures\/Saved Pictures\/angelo koo\/'+date+fname+".xls")
         # df.to_excel('C:\/Users\/s1301\/Pictures\/Saved Pictures\/angelo koo\/'+date+fname+".xls")
 
 
@@ -200,18 +196,18 @@ city_lst = ["w+CAIQICIWTmV3IFlvcmssVW5pdGVkIFN0YXRlcw&gws_rd=cr",
             "w+CAIQICIeQ2hpY2FnbyxJbGxpbm9pcyxVbml0ZWQgU3RhdGVz&gws_rd=cr"]
             "w+CAIQICIeQ2hpY2FnbyxJbGxpbm9pcyxVbml0ZWQgU3RhdGVz&gws_rd=cr"]
 
 
 
 
-cursor=db.query('select json from seo_jobs where cust="美東" and plan="形象SEO" order by rand() limit 1')
-for c in cursor:
-    js=json.loads(c['json'])
-    prefix=js['prefix']
-    postfix=js['postfix']
-    domain=js['domain'][0]
-    positive=js['positive']
-    rnd=js['rnd']
-
-kw1=random.choice(positive)
-kw2=random.choice(rnd)
-kw=prefix+" "+kw1+" "+kw2
-city = random.choice(city_lst)
-url = f"https://www.google.com/search?q={kw}&hl=en&gl=us&num=100&uule={city}&gws_rd=cr"
-run_once({'kw':kw,'url':url})
+# cursor=db.query('select json from seo_jobs where cust="美東" and plan="形象SEO" order by rand() limit 1')
+# for c in cursor:
+#     js=json.loads(c['json'])
+#     prefix=js['prefix']
+#     postfix=js['postfix']
+#     domain=js['domain'][0]
+#     positive=js['positive']
+#     rnd=js['rnd']
+#
+# kw1=random.choice(positive)
+# kw2=random.choice(rnd)
+# kw=prefix+" "+kw1+" "+kw2
+# city = random.choice(city_lst)
+url = "https://www.google.com/search?q=angelo+koo&hl=en&gl=us&num=100&uule=w+CAIQICIWTmV3IFlvcmssVW5pdGVkIFN0YXRlcw&gws_rd=cr"
+run_once({'kw':'angelo koo','url':url})

+ 22 - 18
SEO/seo_notify_gs.py

@@ -15,23 +15,26 @@ from bs4 import BeautifulSoup
 import requests
 import requests
 import dataset
 import dataset
 import time
 import time
-import rpyc
 import sys
 import sys
 import docker
 import docker
-# import pymysql
-# pymysql.install_as_MySQLdb()
-from ga_click import main
+import pymysql
+pymysql.install_as_MySQLdb()
+# from ga_click import main
 
 
 headers = {
 headers = {
     "Authorization": "Bearer " + "OZDcq7sVKwr3F6YNLtBF3LuIgpa4Ql9eAnBWeD7sHTJ",
     "Authorization": "Bearer " + "OZDcq7sVKwr3F6YNLtBF3LuIgpa4Ql9eAnBWeD7sHTJ",
     "Content-Type": "application/x-www-form-urlencoded"
     "Content-Type": "application/x-www-form-urlencoded"
 }
 }
 
 
+name = {'taroboba-yuan.com': '清原', 'hhh.com.tw': ['幸福空間', '幸福空間設計師'], 'sctt.com.tw': '仁本',
+        'gs-rack.com': '神助物流', 'gs-rack.com/tag': '神助物流', 'bennis.com.tw': '班尼斯', 'yogoclean.com': '有夠讚',
+        'yesonlineeng.com': '新飛'}
+
 
 
 def send_msg(kw):
 def send_msg(kw):
     params = {"message": kw}
     params = {"message": kw}
     # print('通知結果', params)
     # print('通知結果', params)
-    r = requests.post("https://notify-api.line.me/api/notify",headers=headers, params=params)
+    # r = requests.post("https://notify-api.line.me/api/notify", headers=headers, params=params)
 
 
 
 
 db = dataset.connect('mysql://choozmo:pAssw0rd@db.ptt.cx:3306/seo?charset=utf8mb4')
 db = dataset.connect('mysql://choozmo:pAssw0rd@db.ptt.cx:3306/seo?charset=utf8mb4')
@@ -42,14 +45,16 @@ cursor_sns = db.query(
     'SELECT * FROM seo.sns_log where datediff(now(),dt)=0 order by dt desc')
     'SELECT * FROM seo.sns_log where datediff(now(),dt)=0 order by dt desc')
 result = "\n"
 result = "\n"
 idx = 0
 idx = 0
-d = {'taroboba-yuan.com':0,'yogoclean.com':0,'yesonlineeng.com':0,'lehua.com.tw':0,'sogoodday.com.tw':0,
-'howday.com.tw':0,'vickybrain.com':0,'crclinic072.com':0,'planta.com.tw':0,'misaglobal.org':0}
+d = {'yogoclean.com': 0, 'sogoodday.com.tw': 0, 'vickybrain.com': 0, 'crclinic072.com': 0, 'planta.com.tw': 0,
+     'misaglobal.org': 0,
+     'kuchichun.com': 0, 'reformbio.asia': 0, 'www.heartwealthy.com': 0, 'cliqproducts.com.tw': 0,'www.hujiai-international.com':0,
+     'lisinbeauty.com':0,'simpleprotein.tw':0,'taroboba-yuan.com':0,'kolin.com.tw':0,'yesonlineeng.com':0}
 for c in cursor:
 for c in cursor:
     custom_name = c['domain']
     custom_name = c['domain']
     if custom_name in d:
     if custom_name in d:
-        if custom_name == 'hhh.com.tw' and c['num']==1:
-            d['hhh.com.tw/designer']+=1
-        d[custom_name]+=1
+        if custom_name == 'hhh.com.tw' and c['num'] == 1:
+            d['hhh.com.tw/designer'] += 1
+        d[custom_name] += 1
     idx += 1
     idx += 1
     if idx <= 3:
     if idx <= 3:
         elmt = c['kw']
         elmt = c['kw']
@@ -58,31 +63,30 @@ for c in cursor:
             custom_name = name[domain]
             custom_name = name[domain]
         except:
         except:
             pass
             pass
-        result += "[" + str(c['ranking']) + "]:" + elmt + " "+ f"({custom_name})\n"
+        result += "[" + str(c['ranking']) + "]:" + elmt + " " + f"({custom_name})\n"
 sns_result = ""
 sns_result = ""
-sns_d = {'班尼斯':0}
+sns_d = {'班尼斯': 0}
 for c in cursor_sns:
 for c in cursor_sns:
     elmt = c['kw']
     elmt = c['kw']
     title = c['title']
     title = c['title']
     custom_name = c['client']
     custom_name = c['client']
     if custom_name in sns_d:
     if custom_name in sns_d:
-        sns_d[custom_name]+=1
+        sns_d[custom_name] += 1
     sns_result = f"\n[{c['ranking']}] {elmt} {title}"
     sns_result = f"\n[{c['ranking']}] {elmt} {title}"
-   
-per_total_click =f"清原:{d['taroboba-yuan.com']}\n新飛:{d['yesonlineeng.com']}\n有夠讚:{d['yogoclean.com']}\n樂樺:{d['lehua.com.tw']}\n好晴天:{d['sogoodday.com.tw']}\n好日子:{d['howday.com.tw']}\n薇懠:{d['vickybrain.com']}\n宏康:{d['crclinic072.com']}\nplantA:{d['planta.com.tw']}\nmisa:{d['misaglobal.org']}\n"
+
+per_total_click = f"有夠讚:{d['yogoclean.com']}\n好晴天:{d['sogoodday.com.tw']}\n薇懠:{d['vickybrain.com']}\n宏康:{d['crclinic072.com']}\nplantA:{d['planta.com.tw']}\nmisa:{d['misaglobal.org']}\n辜琦鈞:{d['kuchichun.com']}\n瑞福:{d['reformbio.asia']}\n鑫富:{d['www.heartwealthy.com']}\nCLIQ露營椅:{d['cliqproducts.com.tw']}\n護全:{d['www.hujiai-international.com']}\n美麗馨:{d['lisinbeauty.com']}\n極簡乳清:{d['simpleprotein.tw']}\n清原:{d['taroboba-yuan.com']}\n歌林:{d['kolin.com.tw']}\n新飛:{d['yesonlineeng.com']}\n"
 sns_per_total_click = f"\n"
 sns_per_total_click = f"\n"
 print(per_total_click)
 print(per_total_click)
 print(sns_per_total_click)
 print(sns_per_total_click)
 # print(sns_result)
 # print(sns_result)
 msg_per_total_click = '\n點擊次數總計:\n' + per_total_click
 msg_per_total_click = '\n點擊次數總計:\n' + per_total_click
 msg = "\n機器人執行次數:" + str(idx)
 msg = "\n機器人執行次數:" + str(idx)
-ga_sessions = '\n幸福空間的GA Sessions'+main()['ga:sessions']
+# ga_sessions = '\n幸福空間的GA Sessions' + main()['ga:sessions']
 sns_per_total_click = '\n點擊次數總計:\n' + sns_per_total_click
 sns_per_total_click = '\n點擊次數總計:\n' + sns_per_total_click
 # print(msg_per_total_click + result + sns_per_total_click + sns_result + ga_sessions)
 # print(msg_per_total_click + result + sns_per_total_click + sns_result + ga_sessions)
-send_msg(msg_per_total_click + result + sns_per_total_click + sns_result + ga_sessions)
+send_msg(msg_per_total_click + result + sns_per_total_click + sns_result)
 db.close()
 db.close()
 
 
-
 # schtasks /create /sc minute /mo 30 /sd 2022/05/050 /st 9:00 /et 23:00 /tn "linebat" /tr "C:\tmp\inn_line.bat"
 # schtasks /create /sc minute /mo 30 /sd 2022/05/050 /st 9:00 /et 23:00 /tn "linebat" /tr "C:\tmp\inn_line.bat"
 
 
 SCRIPT_DIR = os.path.dirname(os.path.abspath(__file__))
 SCRIPT_DIR = os.path.dirname(os.path.abspath(__file__))

+ 1 - 1
SEO/sns_rank_report.py

@@ -4,7 +4,7 @@ import pymysql
 pymysql.install_as_MySQLdb()
 pymysql.install_as_MySQLdb()
 
 
 name='superlink'
 name='superlink'
-date = '1219'
+date = '0109'
 
 
 
 
 def day():
 def day():