openshot_video_generator.py 34 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914
  1. from os import listdir
  2. from os.path import isfile, isdir, join
  3. import openshot
  4. import threading
  5. import zhtts
  6. import os
  7. import urllib
  8. from typing import List
  9. import requests
  10. from pydantic import BaseModel
  11. from bs4 import BeautifulSoup
  12. from PIL import Image,ImageDraw,ImageFont
  13. import pyttsx3
  14. import rpyc
  15. import random
  16. import re
  17. import time
  18. import math
  19. import dataset
  20. from datetime import datetime
  21. from gtts import gTTS
  22. import ffmpy
  23. from difflib import SequenceMatcher
  24. import difflib
  25. from autosub import DEFAULT_CONCURRENCY
  26. from autosub import DEFAULT_SUBTITLE_FORMAT
  27. from pytranscriber.control.ctr_main import Ctr_Main
  28. from pytranscriber.control.ctr_autosub import Ctr_Autosub
  29. import multiprocessing
  30. from itertools import groupby
  31. from operator import itemgetter
  32. dir_sound = 'mp3_track/'
  33. dir_photo = 'photo/'
  34. dir_text = 'text_file/'
  35. dir_video = 'video_material/'
  36. dir_title = 'title/'
  37. dir_subtitle = 'subtitle/'
  38. dir_anchor = 'anchor_raw/'
  39. tmp_video_dir = 'tmp_video/'
  40. video_sub_folder = 'ai_anchor_video/'
  41. dir_list = [dir_sound,dir_photo,dir_text,dir_video,dir_title,dir_subtitle,dir_anchor,tmp_video_dir]
  42. def notify_group(msg):
  43. glist=['7vilzohcyQMPLfAMRloUawiTV4vtusZhxv8Czo7AJX8','WekCRfnAirSiSxALiD6gcm0B56EejsoK89zFbIaiZQD','1dbtJHbWVbrooXmQqc4r8OyRWDryjD4TMJ6DiDsdgsX','HOB1kVNgIb81tTB4Ort1BfhVp9GFo6NlToMQg88vEhh']
  44. for gid in glist:
  45. headers = {
  46. "Authorization": "Bearer " + gid,
  47. "Content-Type": "application/x-www-form-urlencoded"
  48. }
  49. params = {"message": msg}
  50. r = requests.post("https://notify-api.line.me/api/notify",headers=headers, params=params)
  51. def cKey(r,g,b,fuzz):
  52. col=openshot.Color()
  53. col.red=openshot.Keyframe(r)
  54. col.green=openshot.Keyframe(g)
  55. col.blue=openshot.Keyframe(b)
  56. return openshot.ChromaKey(col, openshot.Keyframe(fuzz))
  57. def video_photo_clip(vid=None,layer=None, position=None, end=None
  58. ,scale_x=1,scale_y=1,location_x=0,location_y=0,ck=None,audio=True):
  59. clip = openshot.Clip(vid)
  60. clip.Layer(layer)
  61. clip.Position(position)
  62. clip.End(end)
  63. clip.scale_x=openshot.Keyframe(scale_x)
  64. clip.scale_y=openshot.Keyframe(scale_y)
  65. clip.location_x=openshot.Keyframe(location_x)
  66. clip.location_y=openshot.Keyframe(location_y)
  67. if ck!=None:
  68. clip.AddEffect(ck)
  69. if audio==True:
  70. clip.has_audio=openshot.Keyframe(1)
  71. else:
  72. clip.has_audio=openshot.Keyframe(0)
  73. return clip
  74. def listener_progress(string, percent):
  75. True
  76. def myunichchar(unicode_char):
  77. mb_string = unicode_char.encode('big5')
  78. try:
  79. unicode_char = unichr(ord(mb_string[0]) << 8 | ord(mb_string[1]))
  80. except NameError:
  81. unicode_char = chr(mb_string[0] << 8 | mb_string[1])
  82. return unicode_char
  83. def get_url_type(url):
  84. req = urllib.request.Request(url, method='HEAD', headers={'User-Agent': 'Mozilla/5.0'})
  85. r = urllib.request.urlopen(req)
  86. contentType = r.getheader('Content-Type')
  87. return contentType
  88. def make_dir(name_hash):
  89. for direct in dir_list:
  90. if not os.path.isdir(direct):
  91. os.mkdir(direct)
  92. try:
  93. os.mkdir(dir_photo+name_hash)
  94. except FileExistsError:
  95. print("~~~~~~Warning~~~~~~~~~Directory " , dir_photo+name_hash , " already exists")
  96. try:
  97. os.mkdir(dir_text+name_hash)
  98. except FileExistsError:
  99. print("~~~~~~Warning~~~~~~~~~Directory " , dir_text+name_hash , " already exists")
  100. try:
  101. os.mkdir(dir_sound+name_hash)
  102. except FileExistsError:
  103. print("~~~~~~Warning~~~~~~~~~Directory " , dir_sound+name_hash , " already exists")
  104. try:
  105. os.mkdir(dir_anchor+name_hash)
  106. except FileExistsError:
  107. print("~~~~~~Warning~~~~~~~~~Directory " , dir_anchor+name_hash , " already exists")
  108. try:
  109. os.mkdir(dir_subtitle+name_hash)
  110. except FileExistsError:
  111. print("~~~~~~Warning~~~~~~~~~Directory " , dir_subtitle+name_hash , " already exists")
  112. def file_prepare(name, name_hash,text_content,image_urls,multiLang,lang='zh'):
  113. make_dir(name_hash)
  114. img_num = 1
  115. for imgu in image_urls:
  116. if get_url_type(imgu) =='video/mp4':
  117. r=requests.get(imgu)
  118. f=open(dir_photo+name_hash+"/"+str(img_num)+".mp4",'wb')
  119. for chunk in r.iter_content(chunk_size=255):
  120. if chunk:
  121. f.write(chunk)
  122. f.close()
  123. else:
  124. im = Image.open(requests.get(imgu, stream=True).raw)
  125. im= im.convert("RGB")
  126. im.save(dir_photo+name_hash+"/"+str(img_num)+".jpg")
  127. img_num+=1
  128. #save text
  129. txt_idx=0
  130. for txt in text_content:
  131. text_file = open(dir_text+name_hash+"/"+str(txt_idx)+".txt", "w")
  132. text_file.write(txt)
  133. text_file.close()
  134. txt_idx+=1
  135. print("text file made")
  136. #make mp3
  137. txt_idx = 0
  138. for txt in text_content:
  139. if lang!='zh' or multiLang==1:
  140. if lang!='zh':
  141. tts = gTTS(txt)
  142. tts.save(dir_sound+name_hash+"/"+str(txt_idx)+"raw.mp3")
  143. else:
  144. tts = gTTS(txt,lang='zh-tw')
  145. tts.save(dir_sound+name_hash+"/"+str(txt_idx)+"raw.mp3")
  146. #speed up
  147. ff = ffmpy.FFmpeg(inputs={dir_sound+name_hash+"/"+str(txt_idx)+"raw.mp3": None}
  148. , outputs={dir_sound+name_hash+"/"+str(txt_idx)+".mp3": ["-filter:a", "atempo=1.2"]})
  149. ff.run()
  150. os.remove(dir_sound+name_hash+"/"+str(txt_idx)+"raw.mp3")
  151. else:
  152. print('use zhtts')
  153. tts = zhtts.TTS()
  154. tts.text2wav(txt,dir_sound+name_hash+"/"+str(txt_idx)+".mp3")
  155. txt_idx+=1
  156. print("mp3 file made")
  157. #make title as image
  158. txt2image_title(name, dir_title+name_hash+".png",lang)
  159. def txt2image(content, save_target,lang='zh'):
  160. unicode_text = trim_punctuation(content)
  161. font = ''
  162. if lang=='zh':
  163. font = ImageFont.truetype(font="font/DFT_B7.ttc", size=38)
  164. else :
  165. font = ImageFont.truetype(font="font/arial.ttf", size=38)
  166. text_width, text_height = font.getsize(unicode_text)
  167. canvas = Image.new('RGBA', (700, 500), (255, 0, 0, 0) )
  168. draw = ImageDraw.Draw(canvas)
  169. text= unicode_text
  170. draw.text((5,5), text, (255, 255, 0), font)
  171. canvas.save(save_target, "PNG")
  172. def txt2image_title(content, save_target, lang='zh'):
  173. unicode_text = trim_punctuation(content)
  174. font = ''
  175. if lang=='zh':
  176. font = ImageFont.truetype(font="font/DFT_B7.ttc", size=22)
  177. else :
  178. font = ImageFont.truetype(font="font/arial.ttf", size=22)
  179. text_width, text_height = font.getsize(unicode_text)
  180. canvas = Image.new('RGBA', (510, 500), (255, 0, 0, 0) )
  181. draw = ImageDraw.Draw(canvas)
  182. text= unicode_text
  183. draw.text((5,5), text, (17, 41, 167), font)
  184. canvas.save(save_target, "PNG")
  185. def call_anchor(fileName,avatar):
  186. conn = rpyc.classic.connect("192.168.1.105",18812)
  187. ros = conn.modules.os
  188. rsys = conn.modules.sys
  189. fr=open(dir_sound+fileName+".mp3",'rb')# voice
  190. #warning!!! file my be replaced by other process
  191. fw=conn.builtins.open('/tmp/output.mp3','wb')
  192. while True:
  193. b=fr.read(1024)
  194. if b:
  195. fw.write(b)
  196. else:
  197. break
  198. fr.close()
  199. fw.close()
  200. val=random.randint(1000000,9999999)
  201. ros.chdir('/home/jared/to_video')
  202. ros.system('./p'+str(avatar)+'.sh '+str(val)+' &')
  203. while True:
  204. print('waiting...')
  205. if ros.path.exists('/tmp/results/'+str(val)):
  206. break
  207. time.sleep(5)
  208. print('waiting...')
  209. fr=conn.builtins.open('/tmp/results/'+str(val)+'.mp4','rb')
  210. fw=open(dir_anchor+fileName+".mp4",'wb')
  211. while True:
  212. b=fr.read(1024)
  213. if b:
  214. fw.write(b)
  215. else:
  216. break
  217. fr.close()
  218. fw.close()
  219. def syllable_count(word):
  220. word = word.lower()
  221. count = 0
  222. vowels = "aeiouy"
  223. if word[0] in vowels:
  224. count += 1
  225. for index in range(1, len(word)):
  226. if word[index] in vowels and word[index - 1] not in vowels:
  227. count += 1
  228. if word.endswith("e"):
  229. count -= 1
  230. if count == 0:
  231. count += 1
  232. return count
  233. def split_sentence(in_str, maxLen):
  234. re.findall(r'[\u4e00-\u9fff]+', in_str)
  235. zh_idx = []
  236. eng_idx= []
  237. for i in range(len(in_str)):
  238. if in_str[i] > u'\u4e00' and in_str[i] < u'\u9fff':
  239. zh_idx.append(i)
  240. else:
  241. eng_idx.append(i)
  242. space_index = [m.start() for m in re.finditer(' ', in_str)]
  243. for idx in space_index:
  244. eng_idx.remove(idx)
  245. eng_range_list = []
  246. for k, g in groupby(enumerate(eng_idx), lambda ix : ix[0] - ix[1]):
  247. eng_range = list(map(itemgetter(1), g))
  248. eng_range_list.append(eng_range)
  249. total_syllable = 0
  250. for i in range(len(eng_range_list)):
  251. total_syllable += (syllable_count(in_str[eng_range_list[i][0]:eng_range_list[i][-1]+1])+0.5)
  252. for i in range(len(zh_idx)):
  253. total_syllable+=1
  254. #final chchchchchc[en][en][en]
  255. #[en] is a vocabulary dict with occurence of image
  256. zh_eng_idx_list = []
  257. i = 0
  258. while i < len(in_str):
  259. if in_str[i]==' ':
  260. i+=1
  261. if i in zh_idx:
  262. zh_eng_idx_list.append(i)
  263. i+=1
  264. if i in eng_idx:
  265. for ls in eng_range_list:
  266. if i in ls:
  267. zh_eng_idx_list.append(ls)
  268. i = ls[-1]+1
  269. break
  270. zh_eng_dict_list = [{'content':'','time_ratio':0}]
  271. idx = 0
  272. current_len = 0
  273. sen_idx = 0
  274. while idx < len(zh_eng_idx_list):
  275. str_from_idx = ''
  276. sylla_cnt = 1
  277. if type(zh_eng_idx_list[idx])==type([]):
  278. str_from_idx = in_str[zh_eng_idx_list[idx][0]:zh_eng_idx_list[idx][-1]+1]+' '
  279. sylla_cnt = syllable_count(str_from_idx)
  280. else:
  281. str_from_idx = in_str[zh_eng_idx_list[idx]]
  282. if len(zh_eng_dict_list[sen_idx]['content'])+sylla_cnt>=maxLen:
  283. zh_eng_dict_list[sen_idx]['time_ratio'] = current_len/total_syllable
  284. zh_eng_dict_list.append({'content':'','time_ratio':0})
  285. sen_idx+=1
  286. current_len = 0
  287. else:
  288. current_len += sylla_cnt
  289. zh_eng_dict_list[sen_idx]['content'] += str_from_idx
  290. idx+=1
  291. total_ratio = 0
  292. for obj in zh_eng_dict_list:
  293. total_ratio+=obj['time_ratio']
  294. zh_eng_dict_list[-1]['time_ratio'] = 1-total_ratio
  295. return zh_eng_dict_list
  296. def parse_script(file_path,gt_list):
  297. with open(file_path, 'r',encoding="utf-8") as f:
  298. raw_lines = [line.strip() for line in f]
  299. lines = adjustSub_by_text_similarity(gt_list,raw_lines)
  300. #make dict
  301. dict_list = []
  302. for idx in range(len(lines)):
  303. script={}
  304. script['content'] = lines[idx]
  305. time_raw = raw_lines[idx * 4 +1 ].split(' --> ')
  306. start = time_raw[0].split(':')
  307. stop = time_raw[1].split(':')
  308. script['start'] = float(start[0])*3600 + float(start[1])*60 + float(start[2].replace(',','.'))+0.3
  309. script['stop'] = float(stop[0])*3600 + float(stop[1])*60 + float(stop[2].replace(',','.'))-0.3
  310. dict_list.append(script)
  311. #merge duplicated sentences
  312. script_not_dup_list = []
  313. for idx in range(len(dict_list)):
  314. dup_list = []
  315. for idx_inner in range(len(dict_list)):
  316. if dict_list[idx_inner]['content']==dict_list[idx]['content']:
  317. dup_list.append(idx_inner)
  318. for dup_idx in dup_list:
  319. if dup_idx == min(dup_list):
  320. dict_list[dup_idx]['type'] = 'lead_sentence'
  321. else:
  322. dict_list[dup_idx]['type'] = 'duplicated'
  323. dict_list[dup_list[0]]['stop'] = dict_list[dup_list[-1]]['stop']
  324. if dict_list[idx]['type'] == 'lead_sentence':
  325. script_not_dup_list.append(dict_list[idx])
  326. #avoid subtitle overlapping ? Timeline overlapping not found currently
  327. #cut by max length----> eng seperated problem {eng_idx}
  328. #ENG counts, zh counts, space counts
  329. new_idx = 0
  330. splitted_dict = []
  331. for dic in script_not_dup_list:
  332. dic_idx = 0
  333. accumulated_duration = 0
  334. duration = dic['stop']-dic['start']
  335. print(duration)
  336. for sub_dic in split_sentence(dic['content'],13):
  337. new_dic = {}
  338. new_dic['index'] = new_idx
  339. new_idx+=1
  340. ind_duration = duration * sub_dic['time_ratio']*0.8
  341. new_dic['start'] = dic['start'] + accumulated_duration
  342. accumulated_duration += ind_duration
  343. new_dic['content'] = sub_dic['content']
  344. new_dic['duration'] = ind_duration
  345. splitted_dict.append(new_dic)
  346. return splitted_dict
  347. def adjustSub_by_text_similarity(gts,gens_raw):
  348. gens = []
  349. for idx in range(int((len(gens_raw)+1)/4)):
  350. gens.append(gens_raw[idx*4+2])
  351. combine2 = [''.join([i,j]) for i,j in zip(gts, gts[1:])]
  352. combine3 = [''.join([i,j,k]) for i,j,k in zip(gts, gts[1:], gts[2:])]
  353. alls = gts + combine2 + combine3
  354. adjusted = [None]*len(gens)
  355. duplicated_list = []
  356. for idx in range(len(gens)):
  357. match_text = difflib.get_close_matches(gens[idx], alls, cutoff=0.1)
  358. if match_text[0] in duplicated_list:
  359. for mt in match_text:
  360. if mt == adjusted[idx-1] or mt not in duplicated_list:
  361. adjusted[idx] = mt
  362. break
  363. else:
  364. adjusted[idx] = match_text[0]
  365. duplicated_list.append(match_text[0])
  366. return adjusted
  367. def trim_punctuation(s):
  368. pat_block = u'[^\u4e00-\u9fff0-9a-zA-Z]+';
  369. pattern = u'([0-9]+{0}[0-9]+)|{0}'.format(pat_block)
  370. res = re.sub(pattern, lambda x: x.group(1) if x.group(1) else u" " ,s)
  371. return res
  372. def splitter(s):
  373. for sent in re.findall(u'[^!?,。\!\?]+[!? 。\!\?]?', s, flags=re.U):
  374. yield sent
  375. def split_by_pun(s):
  376. res = list(splitter(s))
  377. return res
  378. def split_by_word(s):
  379. slice_size = 3
  380. paragraph_len = len(s)
  381. slice_num = int(math.ceil(paragraph_len/slice_size))
  382. slice_list = []
  383. for n in range(slice_num):
  384. slice_list.append(s[n*slice_size:n*slice_size+slice_size])
  385. return slice_list
  386. def generate_subtitle_image_from_dict(name_hash, sub_dict):
  387. for script in sub_dict:
  388. sv_path = dir_subtitle + name_hash + '/' + str(script['index'])+'.png'
  389. sub = script['content']
  390. txt2image(sub,sv_path)
  391. def generate_subtitle_image(name_hash,text_content):
  392. img_list = [None]*len(text_content)
  393. for idx in range(len(text_content)):
  394. img_list[idx]=[]
  395. senList = split_by_pun(text_content[idx])
  396. for inner_idx in range(len(senList)):
  397. sv_path = dir_subtitle + name_hash +'/'+str(idx)+ str(inner_idx) +'.png'
  398. sub = senList[inner_idx]
  399. txt2image(sub,sv_path)
  400. img_list[idx]+=[{"count":len(sub),"path":sv_path}]
  401. return img_list
  402. def generate_subtitle_image_ENG(name_hash,text_content):
  403. img_list = [None]*len(text_content)
  404. for idx in range(len(text_content)):
  405. sv_path = dir_subtitle + name_hash +'/'+str(idx)+'.png'
  406. sub = text_content[idx]
  407. txt2image(sub, sv_path,lang='eng')
  408. img_list[idx] = sv_path
  409. return img_list
  410. def video_writer_init(path):
  411. w = openshot.FFmpegWriter(path)
  412. w.SetAudioOptions(True, "aac", 44100, 2, openshot.LAYOUT_STEREO, 3000000)
  413. w.SetVideoOptions(True, "libx264", openshot.Fraction(30000, 1000), 1280, 720,
  414. openshot.Fraction(1, 1), False, False, 3000000)
  415. return w
  416. def video_gen(name_hash,name,text_content, image_urls,multiLang,avatar):
  417. file_prepare(name, name_hash, text_content,image_urls,multiLang)
  418. for fname in range(len(text_content)):
  419. call_anchor(name_hash+"/"+str(fname),avatar)
  420. print('called............................................')
  421. ck=cKey(0,254,0,270)
  422. ck_anchor=cKey(0,255,1,320)
  423. t = openshot.Timeline(1280, 720, openshot.Fraction(30000, 1000), 44100, 2, openshot.LAYOUT_STEREO)
  424. t.Open()
  425. main_timer = 0
  426. LOGO_OP = openshot.FFmpegReader(dir_video+"LOGO_OP_4.mp4")
  427. LOGO_OP.Open() # Open the reader
  428. head_duration = LOGO_OP.info.duration
  429. LOGO_OP_clip = video_photo_clip(vid=LOGO_OP,layer=4,position=0,end=head_duration
  430. ,location_y=-0.03,scale_x=0.8,scale_y=0.704)
  431. t.AddClip(LOGO_OP_clip)
  432. bg_head = openshot.FFmpegReader(dir_video+"complete_head_aispokesgirl.mp4")
  433. bg_head.Open()
  434. bg_head_clip = video_photo_clip(vid=bg_head,layer=2,position=0,end=LOGO_OP.info.duration,ck=ck)
  435. t.AddClip(bg_head_clip)
  436. main_timer += head_duration
  437. bg_head.Close()
  438. LOGO_OP.Close()
  439. anchor = openshot.FFmpegReader(dir_anchor+name_hash+"/0.mp4")
  440. anchor.Open()
  441. #anchor_clip = video_photo_clip(vid=anchor,layer=4,scale_x=0.65,scale_y=0.65,
  442. # location_x=0.35,location_y=0.25,position=main_timer, end=anchor.info.duration,ck=ck_anchor,audio=False)
  443. #t.AddClip(anchor_clip)
  444. speech = openshot.FFmpegReader(dir_sound+name_hash+"/0.mp3")
  445. speech.Open()
  446. speech_clip = openshot.Clip(speech)
  447. speech_clip.Position(main_timer)
  448. speech_clip.End(anchor.info.duration)
  449. t.AddClip(speech_clip)
  450. main_timer += anchor.info.duration
  451. anchor.Close()
  452. speech.Close()
  453. LOGO_ED = openshot.FFmpegReader(dir_video+"LOGO_ED.avi")
  454. LOGO_ED.Open()
  455. LOGO_ED_clip = video_photo_clip(vid=LOGO_ED,layer=4,position=main_timer,end=LOGO_ED.info.duration
  456. ,location_x=0.005,location_y=-0.031, scale_x=0.8,scale_y=0.6825)
  457. t.AddClip(LOGO_ED_clip)
  458. main_timer += LOGO_ED.info.duration
  459. LOGO_ED.Close()
  460. bg = openshot.FFmpegReader(dir_video+"complete_double_aispokesgirl.mp4")
  461. bg.Open()
  462. bg_times = math.floor(main_timer/bg.info.duration)
  463. left_time = (main_timer) % bg.info.duration
  464. bg_clip_list = [None] * bg_times
  465. bg_list = [None] * bg_times
  466. bg.Close()
  467. bg_timer = head_duration
  468. for idx in range(bg_times):
  469. bg_list[idx] = openshot.FFmpegReader(dir_video+"complete_double_aispokesgirl.mp4")
  470. bg_list[idx].Open()
  471. bg_clip_list[idx] = video_photo_clip(bg_list[idx],layer=2,position=bg_timer,end=bg_list[idx].info.duration,ck=ck)
  472. t.AddClip(bg_clip_list[idx])
  473. bg_timer += bg_list[idx].info.duration
  474. bg_list[idx].Close()
  475. bg_left = openshot.FFmpegReader(dir_video+"complete_double_aispokesgirl.mp4")
  476. bg_left.Open()
  477. bg_left_clip = video_photo_clip(bg_left,layer=2,position=bg_timer,end=left_time,ck=ck)
  478. t.AddClip(bg_left_clip)
  479. bg_left.Close()
  480. title = openshot.QtImageReader(dir_title+name_hash+".png")
  481. title.Open() # Open the reader
  482. title_clip = video_photo_clip(vid=title, layer=4,location_x=-0.047, location_y=0.801,position=0,end=head_duration+main_timer)
  483. t.AddClip(title_clip)
  484. w = video_writer_init(tmp_video_dir+name_hash+"raw.mp4")
  485. w.Open()
  486. frames = int(t.info.fps)*int(main_timer)
  487. for n in range(frames):
  488. f=t.GetFrame(n)
  489. w.WriteFrame(f)
  490. t.Close()
  491. w.Close()
  492. print(name+"RAW DONE : www.choozmo.com:8168/"+tmp_video_dir+name_hash+"raw.mp4")
  493. #start adding sub
  494. #add sub
  495. Ctr_Autosub.init()
  496. Ctr_Autosub.generate_subtitles(tmp_video_dir+name_hash+"raw.mp4",'zh',listener_progress,output=tmp_video_dir+name_hash+"script.txt",concurrency=DEFAULT_CONCURRENCY,subtitle_file_format=DEFAULT_SUBTITLE_FORMAT)
  497. sub_dict = parse_script(tmp_video_dir+name_hash+"script.txt",split_by_pun(text_content[0]))
  498. for subd in sub_dict:
  499. print(subd)
  500. generate_subtitle_image_from_dict(name_hash, sub_dict)
  501. #sv_path = dir_subtitle + name_hash + '/' + str(script['index'])+'.png'
  502. t = openshot.Timeline(1280, 720, openshot.Fraction(30000, 1000), 44100, 2, openshot.LAYOUT_STEREO)
  503. t.Open()
  504. raw = openshot.FFmpegReader(tmp_video_dir+name_hash+"raw.mp4")
  505. raw.Open()
  506. raw_clip = video_photo_clip(vid=raw,layer=2,position=0, end=raw.info.duration)
  507. t.AddClip(raw_clip)
  508. sub_img_list = [None] * len(sub_dict)
  509. sub_clip_list = [None] * len(sub_dict)
  510. for sub_obj in sub_dict:
  511. idx = int(sub_obj['index'])
  512. sub_img_list[idx] = openshot.QtImageReader(dir_subtitle + name_hash + '/' + str(idx)+'.png')
  513. sub_img_list[idx].Open()
  514. sub_clip_list[idx] = video_photo_clip(vid=sub_img_list[idx], layer=6,location_x=0.069, location_y=0.89,position=sub_obj['start'],end=math.ceil(sub_obj['duration']))
  515. t.AddClip(sub_clip_list[idx])
  516. sub_img_list[idx].Close()
  517. anchor = openshot.FFmpegReader(dir_anchor+name_hash+"/0.mp4")
  518. anchor.Open()
  519. anchor_clip = video_photo_clip(vid=anchor,layer=4,scale_x=0.65,scale_y=0.65,
  520. location_x=0.35,location_y=0.25,position=head_duration, end=anchor.info.duration,ck=ck_anchor,audio=False)
  521. t.AddClip(anchor_clip)
  522. w = video_writer_init(tmp_video_dir+name_hash+".mp4")
  523. w.Open()
  524. frames = int(t.info.fps)*int(main_timer)
  525. for n in range(frames):
  526. f=t.GetFrame(n)
  527. w.WriteFrame(f)
  528. t.Close()
  529. w.Close()
  530. os.remove(tmp_video_dir+name_hash+"raw.mp4")
  531. os.remove(tmp_video_dir+name_hash+"script.txt")
  532. print(name+"ALL DONE : www.choozmo.com:8168/"+video_sub_folder+name_hash+"raw.mp4")
  533. def anchor_video_v2(name_hash,name,text_content, image_urls,multiLang,avatar):
  534. print(os.getcwd())
  535. print('sub image made')
  536. print(multiLang)
  537. file_prepare(name, name_hash, text_content,image_urls,multiLang)
  538. sub_list=generate_subtitle_image(name_hash,text_content)
  539. for fname in range(len(text_content)):
  540. call_anchor(name_hash+"/"+str(fname),avatar)
  541. print('step finish')
  542. print('called............................................')
  543. ck=cKey(0,254,0,270)
  544. ck_anchor=cKey(0,255,1,320)
  545. duration = 0
  546. #average layer level is 3
  547. t = openshot.Timeline(1280, 720, openshot.Fraction(30000, 1000), 44100, 2, openshot.LAYOUT_STEREO)
  548. t.Open()
  549. main_timer = 0
  550. LOGO_OP = openshot.FFmpegReader(dir_video+"LOGO_OP_4.mp4")
  551. LOGO_OP.Open() # Open the reader
  552. LOGO_OP_clip = video_photo_clip(vid=LOGO_OP,layer=4,position=0,end=LOGO_OP.info.duration
  553. ,location_y=-0.03,scale_x=0.8,scale_y=0.704)
  554. t.AddClip(LOGO_OP_clip)
  555. bg_head = openshot.FFmpegReader(dir_video+"complete_head_aispokesgirl.mp4")
  556. bg_head.Open()
  557. bg_head_clip = video_photo_clip(vid=bg_head,layer=2,position=0,end=LOGO_OP.info.duration,ck=ck)
  558. t.AddClip(bg_head_clip)
  559. main_timer += LOGO_OP.info.duration
  560. head_duration = LOGO_OP.info.duration
  561. bg_head.Close()
  562. LOGO_OP.Close()
  563. clip_duration=0
  564. photo_clip_list = [None]*len(text_content)
  565. img_list = [None]*len(text_content)
  566. anchor_clip_list = [None] * len(text_content)
  567. anchor_list = [None] * len(text_content)
  568. audio_clip_list = [None] * len(text_content)
  569. audio_list = [None] * len(text_content)
  570. sub_clip_list = [None] * len(text_content)
  571. sub_img_list = [None] * len(text_content)
  572. idx = 0
  573. for p in listdir(dir_photo+name_hash):
  574. anchor_list[idx] = openshot.FFmpegReader(dir_anchor+name_hash+"/"+str(idx)+".mp4")
  575. clip_duration = anchor_list[idx].info.duration
  576. anchor_list[idx].Open()
  577. anchor_clip_list[idx] = video_photo_clip(vid=anchor_list[idx],layer=4,scale_x=0.65,scale_y=0.65,
  578. location_x=0.35,location_y=0.25,position=main_timer, end=clip_duration,ck=ck_anchor,audio=False)
  579. t.AddClip(anchor_clip_list[idx])
  580. img_list[idx] = openshot.FFmpegReader(dir_photo+name_hash+'/'+p)
  581. img_list[idx].Open()
  582. photo_clip_list[idx] = video_photo_clip(vid=img_list[idx],layer=3
  583. ,scale_x=0.81,scale_y=0.68,location_y=-0.03,position=main_timer,end=clip_duration,audio=False)
  584. t.AddClip(photo_clip_list[idx])
  585. img_list[idx].Close()
  586. audio_list[idx] = openshot.FFmpegReader(dir_sound+name_hash+"/"+str(idx)+".mp3")
  587. audio_list[idx].Open()
  588. audio_clip_list[idx] = openshot.Clip(audio_list[idx])
  589. audio_clip_list[idx].Position(main_timer)
  590. audio_clip_list[idx].End(clip_duration)
  591. t.AddClip(audio_clip_list[idx])
  592. img_list[idx].Close()
  593. anchor_list[idx].Close()
  594. audio_list[idx].Close()
  595. sub_img_list[idx] = [None] * len(sub_list[idx])
  596. sub_clip_list[idx] = [None] * len(sub_list[idx])
  597. sub_timer = 0
  598. for sub_idx in range(len(sub_list[idx])):
  599. sub_img_list[idx][sub_idx] = openshot.QtImageReader(sub_list[idx][sub_idx]['path'])
  600. sub_img_list[idx][sub_idx].Open()
  601. sub_duration = 0.205*sub_list[idx][sub_idx]['count']
  602. sub_clip_list[idx][sub_idx] = video_photo_clip(vid=sub_img_list[idx][sub_idx], layer=6,location_x=0.069, location_y=0.89,position=main_timer+sub_timer,end=sub_duration)
  603. t.AddClip(sub_clip_list[idx][sub_idx])
  604. sub_img_list[idx][sub_idx].Close()
  605. sub_timer += sub_duration
  606. print(sub_list[idx][sub_idx]['path'])
  607. main_timer += clip_duration
  608. idx+=1
  609. LOGO_ED = openshot.FFmpegReader(dir_video+"LOGO_ED.avi")
  610. LOGO_ED.Open()
  611. LOGO_ED_clip = video_photo_clip(vid=LOGO_ED,layer=4,position=main_timer,end=LOGO_ED.info.duration+2
  612. ,location_x=0.005,location_y=-0.031
  613. ,scale_x=0.8,scale_y=0.6825)
  614. t.AddClip(LOGO_ED_clip)
  615. ED_duration = LOGO_ED.info.duration
  616. LOGO_ED.Close()
  617. bg = openshot.FFmpegReader(dir_video+"complete_double_aispokesgirl.mp4")
  618. bg.Open()
  619. bg_times = math.floor(main_timer+ED_duration/bg.info.duration)
  620. left_time = (main_timer+ED_duration) % bg.info.duration
  621. bg_clip_list = [None] * bg_times
  622. bg_list = [None] * bg_times
  623. bg.Close()
  624. bg_timer = head_duration
  625. for idx in range(bg_times):
  626. bg_list[idx] = openshot.FFmpegReader(dir_video+"complete_double_aispokesgirl.mp4")
  627. bg_list[idx].Open()
  628. bg_clip_list[idx] = video_photo_clip(bg_list[idx],layer=2,position=bg_timer
  629. ,end=bg_list[idx].info.duration,ck=ck)
  630. t.AddClip(bg_clip_list[idx])
  631. bg_timer += bg_list[idx].info.duration
  632. bg_list[idx].Close()
  633. bg_left = openshot.FFmpegReader(dir_video+"complete_double_aispokesgirl.mp4")
  634. bg_left.Open()
  635. bg_left_clip = video_photo_clip(bg_left,layer=2,position=bg_timer,end=left_time,ck=ck)
  636. t.AddClip(bg_left_clip)
  637. bg_left.Close()
  638. title = openshot.QtImageReader(dir_title+name_hash+".png")
  639. title.Open() # Open the reader
  640. title_clip = video_photo_clip(vid=title, layer=4,location_x=-0.047, location_y=0.801,position=0,end=head_duration+main_timer)
  641. t.AddClip(title_clip)
  642. ####start building
  643. w = openshot.FFmpegWriter(tmp_video_dir+name_hash+".mp4")
  644. w.SetAudioOptions(True, "aac", 44100, 2, openshot.LAYOUT_STEREO, 3000000)
  645. w.SetVideoOptions(True, "libx264", openshot.Fraction(30000, 1000), 1280, 720,
  646. openshot.Fraction(1, 1), False, False, 3000000)
  647. w.Open()
  648. #may change duration into t.info.duration
  649. frames = int(t.info.fps)*int(head_duration+main_timer+ED_duration)
  650. for n in range(frames):
  651. f=t.GetFrame(n)
  652. w.WriteFrame(f)
  653. #notify_group(name+"的影片已經產生完成囉! www.choozmo.com:8168/"+video_sub_folder+name_hash+".mp4")
  654. t.Close()
  655. w.Close()
  656. print("video at : www.choozmo.com:8168/"+video_sub_folder+name_hash+".mp4")
  657. def anchor_video_eng(name_hash,name,text_content, image_urls,sub_titles,avatar):
  658. file_prepare(name, name_hash, text_content,image_urls,'eng')
  659. sub_list=generate_subtitle_image_ENG(name_hash,sub_titles)
  660. for fname in range(len(text_content)):
  661. call_anchor(name_hash+"/"+str(fname),avatar)
  662. print('step finish')
  663. print('called............................................')
  664. ck=cKey(0,254,0,270)
  665. ck_anchor=cKey(0,255,1,320)
  666. duration = 0
  667. #average layer level is 3
  668. t = openshot.Timeline(1280, 720, openshot.Fraction(30000, 1000), 44100, 2, openshot.LAYOUT_STEREO)
  669. t.Open()
  670. main_timer = 0
  671. #add logo
  672. LOGO_OP = openshot.FFmpegReader(dir_video+"LOGO_OP_4.mp4")
  673. LOGO_OP.Open() # Open the reader
  674. LOGO_OP_clip = video_photo_clip(vid=LOGO_OP,layer=4,position=0,end=LOGO_OP.info.duration
  675. ,location_y=-0.03,scale_x=0.8,scale_y=0.704)
  676. t.AddClip(LOGO_OP_clip)
  677. #add background video (head is different)
  678. bg_head = openshot.FFmpegReader(dir_video+"complete_head_aispokesgirl.mp4")
  679. bg_head.Open()
  680. bg_head_clip = video_photo_clip(vid=bg_head,layer=2,position=0,end=LOGO_OP.info.duration,ck=ck)
  681. t.AddClip(bg_head_clip)
  682. main_timer += LOGO_OP.info.duration
  683. head_duration = LOGO_OP.info.duration
  684. bg_head.Close()
  685. LOGO_OP.Close()
  686. #prepare empty list
  687. clip_duration=0
  688. photo_clip_list = [None]*len(text_content)
  689. img_list = [None]*len(text_content)
  690. anchor_clip_list = [None] * len(text_content)
  691. anchor_list = [None] * len(text_content)
  692. audio_clip_list = [None] * len(text_content)
  693. audio_list = [None] * len(text_content)
  694. sub_clip_list = [None] * len(text_content)
  695. #openshot image holder
  696. sub_img_list = [None] * len(text_content)
  697. idx = 0
  698. for p in listdir(dir_photo+name_hash):
  699. anchor_list[idx] = openshot.FFmpegReader(dir_anchor+name_hash+"/"+str(idx)+".mp4")
  700. clip_duration = anchor_list[idx].info.duration
  701. anchor_list[idx].Open()
  702. anchor_clip_list[idx] = video_photo_clip(vid=anchor_list[idx],layer=4,scale_x=0.65,scale_y=0.65,
  703. location_x=0.35,location_y=0.25,position=main_timer, end=clip_duration,ck=ck_anchor,audio=False)
  704. t.AddClip(anchor_clip_list[idx])
  705. #insert image
  706. img_list[idx] = openshot.FFmpegReader(dir_photo+name_hash+'/'+p)
  707. img_list[idx].Open()
  708. photo_clip_list[idx] = video_photo_clip(vid=img_list[idx],layer=3
  709. ,scale_x=0.81,scale_y=0.68,location_y=-0.03,position=main_timer,end=clip_duration,audio=False)
  710. t.AddClip(photo_clip_list[idx])
  711. img_list[idx].Close()
  712. #insert audio (speech)
  713. audio_list[idx] = openshot.FFmpegReader(dir_sound+name_hash+"/"+str(idx)+".mp3")
  714. audio_list[idx].Open()
  715. audio_clip_list[idx] = openshot.Clip(audio_list[idx])
  716. audio_clip_list[idx].Position(main_timer)
  717. audio_clip_list[idx].End(clip_duration)
  718. t.AddClip(audio_clip_list[idx])
  719. #insert subtitle
  720. sub_img_list[idx] = openshot.QtImageReader(sub_list[idx])
  721. sub_img_list[idx].Open()
  722. sub_clip_list[idx] = video_photo_clip(vid=sub_img_list[idx], layer=6,location_x=0.069, location_y=0.89,position=main_timer,end=clip_duration)
  723. t.AddClip(sub_clip_list[idx])
  724. img_list[idx].Close()
  725. anchor_list[idx].Close()
  726. audio_list[idx].Close()
  727. sub_img_list[idx].Close()
  728. main_timer += clip_duration
  729. idx+=1
  730. LOGO_ED = openshot.FFmpegReader(dir_video+"ED_ENG.mp4")
  731. LOGO_ED.Open()
  732. LOGO_ED_clip = video_photo_clip(vid=LOGO_ED,layer=4,position=main_timer,end=LOGO_ED.info.duration+2
  733. ,location_x=0.005,location_y=-0.031
  734. ,scale_x=0.8,scale_y=0.6825)
  735. t.AddClip(LOGO_ED_clip)
  736. ED_duration = LOGO_ED.info.duration
  737. LOGO_ED.Close()
  738. bg = openshot.FFmpegReader(dir_video+"complete_double_aispokesgirl.mp4")
  739. bg.Open()
  740. bg_times = math.floor(main_timer+ED_duration/bg.info.duration)
  741. left_time = (main_timer+ED_duration) % bg.info.duration
  742. bg_clip_list = [None] * bg_times
  743. bg_list = [None] * bg_times
  744. bg.Close()
  745. bg_timer = head_duration
  746. for idx in range(bg_times):
  747. bg_list[idx] = openshot.FFmpegReader(dir_video+"complete_double_aispokesgirl.mp4")
  748. bg_list[idx].Open()
  749. bg_clip_list[idx] = video_photo_clip(bg_list[idx],layer=2,position=bg_timer
  750. ,end=bg_list[idx].info.duration,ck=ck)
  751. t.AddClip(bg_clip_list[idx])
  752. bg_timer += bg_list[idx].info.duration
  753. bg_list[idx].Close()
  754. bg_left = openshot.FFmpegReader(dir_video+"complete_double_aispokesgirl.mp4")
  755. bg_left.Open()
  756. bg_left_clip = video_photo_clip(bg_left,layer=2,position=bg_timer,end=left_time,ck=ck)
  757. t.AddClip(bg_left_clip)
  758. bg_left.Close()
  759. title = openshot.QtImageReader(dir_title+name_hash+".png")
  760. title.Open() # Open the reader
  761. title_clip = video_photo_clip(vid=title, layer=4,location_x=-0.047, location_y=0.801,position=0,end=head_duration+main_timer)
  762. t.AddClip(title_clip)
  763. ####start building
  764. w = openshot.FFmpegWriter(tmp_video_dir+name_hash+".mp4")
  765. w.SetAudioOptions(True, "aac", 44100, 2, openshot.LAYOUT_STEREO, 3000000)
  766. w.SetVideoOptions(True, "libx264", openshot.Fraction(30000, 1000), 1280, 720,
  767. openshot.Fraction(1, 1), False, False, 3000000)
  768. w.Open()
  769. #may change duration into t.info.duration
  770. frames = int(t.info.fps)*int(head_duration+main_timer+ED_duration)
  771. for n in range(frames):
  772. f=t.GetFrame(n)
  773. w.WriteFrame(f)
  774. #notify_group(name+"(ENG)的影片已經產生完成囉! www.choozmo.com:8168/"+video_sub_folder+name_hash+".mp4")
  775. t.Close()
  776. w.Close()
  777. print("video at : www.choozmo.com:8168/"+video_sub_folder+name_hash+".mp4")
  778. #line notifs
  779. import pyttsx3
  780. def make_speech(text):
  781. engine = pyttsx3.init()
  782. #voices = engine.getProperty('voices')
  783. engine.setProperty('voice', 'Mandarin')
  784. engine.save_to_file(text, '/app/speech.mp3')
  785. engine.runAndWait()
  786. class video_service(rpyc.Service):
  787. def exposed_call_video(self,name_hash,name,text_content, image_urls,multiLang,avatar):
  788. print('ML:'+str(multiLang))
  789. anchor_video_v2(name_hash,name,text_content, image_urls,multiLang,avatar)
  790. def exposed_call_video_eng(self,name_hash,name,text_content, image_urls,sub_titles,avatar):
  791. anchor_video_eng(name_hash,name,text_content, image_urls,sub_titles,avatar)
  792. def exposed_call_video_gen(self,name_hash,name,text_content, image_urls,multiLang,avatar):
  793. print('ML:'+str(multiLang))#this is long video version,
  794. video_gen(name_hash,name,text_content, image_urls,multiLang,avatar)
  795. def exposed_make_speech(self,text):
  796. make_speech(text)
  797. from rpyc.utils.server import ThreadedServer
  798. t = ThreadedServer(video_service, port=8858)
  799. print('service started')
  800. t.start()