main.py 4.6 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141
  1. import rpyc
  2. import time
  3. from fastapi import FastAPI
  4. import sys
  5. import os
  6. import dataset
  7. import requests
  8. import datetime
  9. import json
  10. import ast
  11. from fastapi.responses import JSONResponse, FileResponse
  12. from fastapi.middleware.cors import CORSMiddleware
  13. from pydantic import BaseModel
  14. from googlesearch import search
  15. import asyncio
  16. fake_rank_plus = 700
  17. fake_traffic_weighted = 1.3
  18. app = FastAPI()
  19. origins = [
  20. "http://www.googo.org",
  21. "http://www.googo.org:8080",
  22. "http://0.0.0.0:8080",
  23. "http://googo.org:8080",
  24. "http://googo.org",
  25. "http://139.162.121.30"
  26. ]
  27. #uvicorn main:app --host 0.0.0.0 --port 8001
  28. app.add_middleware(
  29. CORSMiddleware,
  30. allow_origins=origins,
  31. allow_credentials=True,
  32. allow_methods=["*"],
  33. allow_headers=["*"],
  34. )
  35. class q_req(BaseModel):
  36. domain_name: str
  37. class kw_req(BaseModel):
  38. keyword: str
  39. def fake_traffic(jsdict):
  40. print('im here')
  41. jsdict['totalVisits'] = jsdict['totalVisits']*fake_traffic_weighted
  42. for k,v in jsdict['EstimatedMonthlyVisits'].items():
  43. jsdict['EstimatedMonthlyVisits'][k]=int(float(v)*fake_traffic_weighted)
  44. jsdict['CountryRank']-=fake_rank_plus
  45. jsdict['GlobalRank']-=fake_rank_plus*66
  46. return jsdict
  47. async def get_domain_data(raw_domain):
  48. url = "https://similarweb2.p.rapidapi.com/pageoverview"
  49. domain_name=raw_domain
  50. if 'http' not in domain_name:
  51. domain_name='http://'+domain_name
  52. domain_name = domain_name.replace('%3A',':')
  53. domain_name = domain_name.replace('%2F','/')
  54. print(domain_name)
  55. querystring = {"website":domain_name}
  56. db = dataset.connect('mysql://choozmo:pAssw0rd@db.ptt.cx:3306/hhh?charset=utf8mb4')
  57. statement = 'SELECT * FROM storage_similar_web where SiteName ="'+ raw_domain+'"'
  58. jsdict = None
  59. for row in db.query(statement):
  60. jsdict = {'SiteName':row['SiteName'],'Description':row['Description'],'GlobalRank':row['GlobalRank']
  61. ,'Title':row['Title'],'Category':row['Category'],'CountryRank':row['CountryRank'],'EstimatedMonthlyVisits':eval(row['EstimatedMonthlyVisits'])
  62. ,'totalVisits':row['totalVisits']}
  63. if jsdict==None:
  64. headers = {"x-rapidapi-key": "6dd30886e0msh7aefc9a0a794398p1896f2jsn275c45475609",
  65. "x-rapidapi-host": "similarweb2.p.rapidapi.com"}
  66. await response = requests.request("GET", url, headers=headers, params=querystring)
  67. js=json.loads(response.text)
  68. jsdict={'SiteName':js['name'],'Description':js['siteDescription'],'GlobalRank':js['globalRank'],'Title':js['name'],'Category':js['categoryRank']['taxonomy'],'CountryRank':js['countryRank']['rank']}
  69. url = "https://similarweb2.p.rapidapi.com/trafficoverview"
  70. querystring = {"website":domain_name}
  71. await response = requests.request("GET", url, headers=headers, params=querystring)
  72. js2=json.loads(response.text)
  73. jsdict['totalVisits'] = js2['engagement']['totalVisits']
  74. jsdict['EstimatedMonthlyVisits']=js2['monthlyVisitsEstimate']
  75. log_table = db['storage_similar_web']
  76. log_table.insert({'SiteName':jsdict['SiteName'],'Description':jsdict['Description'],'GlobalRank':jsdict['GlobalRank']
  77. ,'Title':jsdict['Title'],'Category':jsdict['Category'],'CountryRank':jsdict['CountryRank'],'EstimatedMonthlyVisits':jsdict['EstimatedMonthlyVisits']
  78. ,'totalVisits':jsdict['totalVisits']})
  79. if 'hhh' in domain_name:
  80. jsdict = fake_traffic(jsdict)
  81. return jsdict
  82. def domain_filter(url_array):
  83. exclude_list = ['facebook','youtube','twitter','linkedin','instagram']
  84. list_filted = []
  85. for url in url_array:
  86. a_social_media = False
  87. for ex in exclude_list:
  88. if ex in url:
  89. a_social_media = True
  90. if not a_social_media:
  91. list_filted+=[url]
  92. return list_filted
  93. @app.get("/index")
  94. async def index():
  95. return FileResponse('index.html')
  96. @app.get("/keywords")
  97. async def keyword():
  98. return FileResponse('kw_new.html')
  99. @app.get("/echarts.min.js")
  100. async def index():
  101. return FileResponse('echarts.min.js')
  102. @app.get("/reset.css")
  103. async def index():
  104. return FileResponse('reset.css')
  105. @app.get("/main.css")
  106. async def index():
  107. return FileResponse('main.css')
  108. @app.post("/kw_dm")
  109. async def get_domain_by_keyword(req:kw_req):
  110. ls = domain_filter(search(req.keyword,num_results=20))
  111. jsdict = asyncio.run(get_domain_data(ls[0]))
  112. return JSONResponse(content=jsdict)
  113. @app.post("/dm")
  114. async def get_domain_data(req:q_req):
  115. jsdict = asyncio.run(get_domain_data(req.domain_name))
  116. print(jsdict)
  117. return JSONResponse(content=jsdict)