Compare commits
No commits in common. "master" and "f3b04f9459dce9a8d34c55a6752807a2dde49a12" have entirely different histories.
master
...
f3b04f9459
1
.gitignore
vendored
1
.gitignore
vendored
@ -2,4 +2,3 @@ backup*/
|
|||||||
wp-navigation
|
wp-navigation
|
||||||
*.log
|
*.log
|
||||||
__pycache__/
|
__pycache__/
|
||||||
wp-gallery
|
|
@ -3,35 +3,15 @@ from requests.auth import HTTPBasicAuth
|
|||||||
from getpass import getpass
|
from getpass import getpass
|
||||||
from urllib.parse import urlparse
|
from urllib.parse import urlparse
|
||||||
from concurrent import futures
|
from concurrent import futures
|
||||||
from concurrent.futures import as_completed, wait, ALL_COMPLETED
|
from concurrent.futures import as_completed, wait
|
||||||
|
|
||||||
import argparse, logging, threading, os, glob
|
import argparse, logging, threading
|
||||||
from lib.WPImport import WPimport
|
from lib.WPImport import WPimport
|
||||||
from lib.WPExport import WPExport
|
from lib.WPExport import WPExport
|
||||||
from lib.WPRemove import WPRemove
|
from lib.WPRemove import WPRemove
|
||||||
from lib.WPChange import WPChange
|
|
||||||
from lib.WPMenu import WPMenu
|
|
||||||
|
|
||||||
def errorRevert(logger, revert, tmp):
|
def remove(args, basic, logger):
|
||||||
if revert is True:
|
removeWp = WPRemove(basic=basic, wordpress="", logger=logger)
|
||||||
files_tmp = glob.glob("{0}/*.json".format(tmp))
|
|
||||||
if len(files_tmp) == 0:
|
|
||||||
logger.error("Error revert, because files not found")
|
|
||||||
exit(1)
|
|
||||||
if len(files_tmp) != int(args.parallel):
|
|
||||||
for file_r in files_tmp:
|
|
||||||
os.remove(file_r)
|
|
||||||
logger.error("Error revert, because number files tmp is incompatible with parallel number")
|
|
||||||
exit(1)
|
|
||||||
|
|
||||||
def change(index, number, args, logger, tmp, revert):
|
|
||||||
changeWp = WPChange(logger=logger, index_name=index, number_thread=number, tmp=tmp)
|
|
||||||
changeWp.fromDirectory(args.directory, revert)
|
|
||||||
|
|
||||||
del changeWp
|
|
||||||
|
|
||||||
def remove(index, number, args, basic, logger, ssl_wordpress):
|
|
||||||
removeWp = WPRemove(basic=basic, wordpress="", logger=logger, ssl_wordpress=ssl_wordpress, index_name=index, number_thread=number)
|
|
||||||
if args.remove == True:
|
if args.remove == True:
|
||||||
for i in args.wordpress.split(","):
|
for i in args.wordpress.split(","):
|
||||||
removeWp.setUrl(i)
|
removeWp.setUrl(i)
|
||||||
@ -53,45 +33,42 @@ def remove(index, number, args, basic, logger, ssl_wordpress):
|
|||||||
del removeWp
|
del removeWp
|
||||||
|
|
||||||
|
|
||||||
def download(name_thread, max_thread, url, logger, parser, directory, html, img, ssl_canalblog, revert, tmp):
|
def download(name_thread, max_thread, url, logger, parser, directory, html, img):
|
||||||
exportWp = WPExport(name="Thread-{0}".format(int(name_thread) + 1), url=url, logger=logger, parser=parser, directory=directory, ssl_canalblog=ssl_canalblog)
|
exportWp = WPExport(name="Thread-{0}".format(int(name_thread) + 1), url=url, logger=logger, parser=parser, directory=directory)
|
||||||
if revert is False:
|
webpage = exportWp.getUrlPage(name_thread, max_thread)
|
||||||
exportWp.getUrlPage(name_thread, max_thread)
|
|
||||||
for i in ["article", "page"]:
|
for i in ["article", "page"]:
|
||||||
for j in ["publications", "principal"]:
|
for j in ["publications", "principal"]:
|
||||||
if html is False:
|
if html is False:
|
||||||
exportWp.downloadHTML(j, i)
|
exportWp.downloadHTML(webpage[j][i])
|
||||||
|
|
||||||
if img is False:
|
if img is False:
|
||||||
exportWp.downloadImg(j, i)
|
exportWp.downloadImg(webpage[j][i])
|
||||||
del exportWp
|
|
||||||
|
|
||||||
|
|
||||||
def importUrl(name_thread, max_thread, canalblog, logger, parser, wordpress, basic, serial, ssl_wordpress, ssl_canalblog, create, update, image, revert, tmp, author):
|
|
||||||
|
|
||||||
|
def importUrl(name_thread, max_thread, canalblog, logger, parser, wordpress, basic, serial):
|
||||||
canalblog = canalblog.split(",")
|
canalblog = canalblog.split(",")
|
||||||
wordpress = wordpress.split(",")
|
wordpress = wordpress.split(",")
|
||||||
name = "Thread-{0}".format(int(name_thread) + 1)
|
name = "Thread-{0}".format(int(name_thread) + 1)
|
||||||
protocol = "https"
|
|
||||||
if ssl_canalblog is False:
|
|
||||||
protocol = "http"
|
|
||||||
if serial is False:
|
if serial is False:
|
||||||
for canal in canalblog:
|
for canal in canalblog:
|
||||||
try:
|
try:
|
||||||
o = urlparse(canal)
|
o = urlparse(canal)
|
||||||
o = o._replace(scheme=protocol)
|
o = o._replace(scheme="https")
|
||||||
url = o.geturl().replace(":///", "://")
|
url = o.geturl().replace(":///", "://")
|
||||||
except Exception as err:
|
except Exception as err:
|
||||||
logger.error("{0} : parsing error : {1}".format(name, err))
|
logger.error("{0} : parsing error : {1}".format(name, err))
|
||||||
exit(1)
|
exit(1)
|
||||||
exportWp = WPExport(name="Thread-{0}".format(int(name_thread) + 1), url=url, logger=logger, parser=parser, ssl_canalblog=ssl_canalblog, tmp=tmp)
|
exportWp = WPExport(name="Thread-{0}".format(int(name_thread) + 1), url=url, logger=logger, parser=parser)
|
||||||
if not revert:
|
webpage = exportWp.getUrlPage(name_thread, max_thread)
|
||||||
exportWp.getUrlPage(name_thread, max_thread)
|
|
||||||
del exportWp
|
del exportWp
|
||||||
for j in wordpress:
|
for j in wordpress:
|
||||||
importWp = WPimport(name=name, basic=basic, wordpress=j, logger=logger, parser=parser, ssl_wordpress=ssl_wordpress, no_create=create, no_update=update, no_image=image, tmp=tmp, author=author)
|
importWp = WPimport(name=name, basic=basic, wordpress=j, logger=logger, parser=parser)
|
||||||
for k in ["article", "page"]:
|
for k in ["article", "page"]:
|
||||||
for l in ["publications", "principal"]:
|
for l in ["publications", "principal"]:
|
||||||
importWp.fromUrl(l, k)
|
importWp.fromUrl(webpage[l][k])
|
||||||
|
|
||||||
del importWp
|
del importWp
|
||||||
else:
|
else:
|
||||||
@ -101,16 +78,15 @@ def importUrl(name_thread, max_thread, canalblog, logger, parser, wordpress, bas
|
|||||||
for i in range(0, len(canalblog)-1):
|
for i in range(0, len(canalblog)-1):
|
||||||
try:
|
try:
|
||||||
o = urlparse(canalblog[i])
|
o = urlparse(canalblog[i])
|
||||||
o = o._replace(scheme=protocol)
|
o = o._replace(scheme="https")
|
||||||
url = o.geturl().replace(":///", "://")
|
url = o.geturl().replace(":///", "://")
|
||||||
except Exception as err:
|
except Exception as err:
|
||||||
logger.error("parsing error : {0}".format(err))
|
logger.error("parsing error : {0}".format(err))
|
||||||
exit(1)
|
exit(1)
|
||||||
exportWp = WPExport(name=name, url=url, logger=logger, parser=parser, ssl_canalblog=ssl_canalblog)
|
exportWp = WPExport(name=name, url=url, logger=logger, parser=parser)
|
||||||
if not revert:
|
webpage = exportWp.getUrlPage(name_thread, max_thread)
|
||||||
exportWp.getUrlPage(name_thread, max_thread)
|
|
||||||
del exportWp
|
del exportWp
|
||||||
importWp = WPimport(name=name, basic=basic, wordpress=wordpress[i], logger=logger, parser=parser, ssl_wordpress=ssl_wordpress, no_create=create, no_update=update, no_image=image, tmp=tmp, author=author)
|
importWp = WPimport(name=name, basic=basic, wordpress=wordpress[i], logger=logger, parser=parser)
|
||||||
|
|
||||||
for k in ["article", "page"]:
|
for k in ["article", "page"]:
|
||||||
for l in ["publications", "principal"]:
|
for l in ["publications", "principal"]:
|
||||||
@ -119,24 +95,24 @@ def importUrl(name_thread, max_thread, canalblog, logger, parser, wordpress, bas
|
|||||||
del importWp
|
del importWp
|
||||||
|
|
||||||
|
|
||||||
def importDirectory(name_thread, max_thread, directory, logger, parser, wordpress, basic, serial, ssl_wordpress, create, update, image, revert, author):
|
def importDirectory(name_thread, max_thread, directory, logger, parser, wordpress, basic, serial):
|
||||||
name = "Thread-{0}".format(int(name_thread) + 1)
|
name = "Thread-{0}".format(int(name_thread) + 1)
|
||||||
directory = directory.split(",")
|
directory = directory.split(",")
|
||||||
wordpress = wordpress.split(",")
|
wordpress = wordpress.split(",")
|
||||||
if serial is False:
|
if serial is False:
|
||||||
for i in wordpress:
|
for i in wordpress:
|
||||||
importWp = WPimport(name=name, basic=basic, wordpress=i, logger=logger, parser=parser, ssl_wordpress=ssl_wordpress, no_create=create, no_update=update, no_image=image, author=author)
|
importWp = WPimport(name=name, basic=basic, wordpress=i, logger=logger, parser=parser)
|
||||||
for j in directory:
|
for j in directory:
|
||||||
importWp.fromDirectory(j, name_thread, max_thread, revert)
|
importWp.fromDirectory(j, name_thread, max_thread)
|
||||||
del importWp
|
del importWp
|
||||||
|
|
||||||
else:
|
else:
|
||||||
if len(directory) != len(wordpress):
|
if len(directory) != len(wordpress):
|
||||||
logger.error("{0} : Error : Number directory is different than wordpress".format(name))
|
logger.error("{0} : Error : Number directory is differant than wordpress".format(name))
|
||||||
exit(1)
|
exit(1)
|
||||||
for i in range(0, len(wordpress)-1):
|
for i in range(0, len(wordpress)-1):
|
||||||
importWp = WPimport(name=name, basic=basic, wordpress=wordpress[i], logger=logger, parser=parser, ssl_wordpress=ssl_wordpress, no_create=create, no_update=update, no_image=image, author=author)
|
importWp = WPimport(name=name, basic=basic, wordpress=wordpress[i], logger=logger, parser=parser)
|
||||||
importWp.fromDirectory(directory[i], name_thread, max_thread, revert)
|
importWp.fromDirectory(directory[i])
|
||||||
del importWp
|
del importWp
|
||||||
|
|
||||||
|
|
||||||
@ -148,9 +124,6 @@ if __name__ == '__main__':
|
|||||||
parser.add_argument("--quiet", help="No console output", action="store_true")
|
parser.add_argument("--quiet", help="No console output", action="store_true")
|
||||||
parser.add_argument("--parser", help="Parser content", default="html.parser")
|
parser.add_argument("--parser", help="Parser content", default="html.parser")
|
||||||
parser.add_argument("--parallel", help="Define number thread (default : 1)", default=1)
|
parser.add_argument("--parallel", help="Define number thread (default : 1)", default=1)
|
||||||
parser.add_argument("--no-ssl", help="No ssl for canalblog and/or wordpress (example wordpress,canalblog)", dest="ssl", default="")
|
|
||||||
parser.add_argument("--revert", help="Restart a work from stopping work", action="store_true")
|
|
||||||
parser.add_argument("--tmp", help="directory tmp", default="/tmp/import_export_canablog")
|
|
||||||
|
|
||||||
subparsers = parser.add_subparsers(dest="command")
|
subparsers = parser.add_subparsers(dest="command")
|
||||||
|
|
||||||
@ -167,13 +140,6 @@ if __name__ == '__main__':
|
|||||||
import_parser.add_argument("--remove-categories", help="Remove all categories", dest="categories", action="store_true")
|
import_parser.add_argument("--remove-categories", help="Remove all categories", dest="categories", action="store_true")
|
||||||
import_parser.add_argument("--remove-tags", help="Remove all tags", dest="tags", action="store_true")
|
import_parser.add_argument("--remove-tags", help="Remove all tags", dest="tags", action="store_true")
|
||||||
import_parser.add_argument("--remove-media", help="Remove all media", dest="media", action="store_true")
|
import_parser.add_argument("--remove-media", help="Remove all media", dest="media", action="store_true")
|
||||||
import_parser.add_argument("--no-create", help="No create post", dest="create", default="store_false", action="store_true")
|
|
||||||
import_parser.add_argument("--no-update", help="No update post", dest="update", default="store_false", action="store_true")
|
|
||||||
import_parser.add_argument("--no-image", help="No image add or update", dest="image", default="store_false", action="store_true")
|
|
||||||
import_parser.add_argument("--no-menu", help="No menu add or update", dest="menu", default="store_false", action="store_true")
|
|
||||||
|
|
||||||
import_parser.add_argument("--author", dest="author", help="Define author", default="")
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
remove_parser = subparsers.add_parser("remove")
|
remove_parser = subparsers.add_parser("remove")
|
||||||
@ -199,33 +165,12 @@ if __name__ == '__main__':
|
|||||||
export_parser.add_argument("--no-img", help="No img", dest="img", action="store_true")
|
export_parser.add_argument("--no-img", help="No img", dest="img", action="store_true")
|
||||||
export_parser.add_argument("--no-html", help="No HTML", dest="html", action="store_true")
|
export_parser.add_argument("--no-html", help="No HTML", dest="html", action="store_true")
|
||||||
|
|
||||||
change_parser = subparsers.add_parser("change")
|
|
||||||
change_parser.add_argument("--directory",
|
|
||||||
default="",
|
|
||||||
help="Directory")
|
|
||||||
change_parser.add_argument("--file",
|
|
||||||
default="",
|
|
||||||
help="File")
|
|
||||||
|
|
||||||
menu_parser = subparsers.add_parser("menu")
|
|
||||||
menu_parser.add_argument("--user", help="wordpress user", required=True)
|
|
||||||
menu_parser.add_argument("--password", help="password wordpress's user", default="")
|
|
||||||
menu_parser.add_argument("--file", help="HTML file", default="")
|
|
||||||
menu_parser.add_argument("--canalblog", help="URL Canalblog", default="")
|
|
||||||
menu_parser.add_argument("--wordpress", help="URL Wordpress", required=True)
|
|
||||||
|
|
||||||
args = parser.parse_args()
|
args = parser.parse_args()
|
||||||
|
|
||||||
logger = logging.getLogger('import export canalblog')
|
logger = logging.getLogger('import export canalblog')
|
||||||
formatter = logging.Formatter('%(asctime)s - %(name)s - %(levelname)s - %(message)s')
|
formatter = logging.Formatter('%(asctime)s - %(name)s - %(levelname)s - %(message)s')
|
||||||
ssl_canalblog = True
|
|
||||||
ssl_wordpress = True
|
|
||||||
|
|
||||||
for i in args.ssl.split(","):
|
|
||||||
if i == "canalblog":
|
|
||||||
ssl_canalblog = False
|
|
||||||
if i == "wordpress":
|
|
||||||
ssl_wordpress = False
|
|
||||||
|
|
||||||
if args.quiet is False:
|
if args.quiet is False:
|
||||||
ch = logging.StreamHandler()
|
ch = logging.StreamHandler()
|
||||||
@ -248,9 +193,7 @@ if __name__ == '__main__':
|
|||||||
fileHandler.setFormatter(formatter)
|
fileHandler.setFormatter(formatter)
|
||||||
logger.addHandler(fileHandler)
|
logger.addHandler(fileHandler)
|
||||||
|
|
||||||
os.makedirs(args.tmp, exist_ok=True)
|
if args.command == "import" or args.command == "remove":
|
||||||
|
|
||||||
if args.command == "import" or args.command == "remove" or args.command == "menu":
|
|
||||||
password = args.password
|
password = args.password
|
||||||
if len(args.password) == 0:
|
if len(args.password) == 0:
|
||||||
password = getpass()
|
password = getpass()
|
||||||
@ -261,47 +204,29 @@ if __name__ == '__main__':
|
|||||||
basic = HTTPBasicAuth(args.user, password)
|
basic = HTTPBasicAuth(args.user, password)
|
||||||
if args.command == "import":
|
if args.command == "import":
|
||||||
wordpress = args.wordpress.split(",")
|
wordpress = args.wordpress.split(",")
|
||||||
importWp = WPimport(basic=basic, wordpress="", logger=logger, parser=args.parser, ssl_wordpress=ssl_wordpress, author=args.author, ssl_canalblog=ssl_canalblog)
|
importWp = WPimport(basic=basic, wordpress="", logger=logger, parser=args.parser)
|
||||||
if len(args.file) > 0:
|
if len(args.file) > 0:
|
||||||
for i in wordpress:
|
for i in wordpress:
|
||||||
importWp.setUrl(i)
|
importWp.setUrl(i)
|
||||||
importWp.fromFile(files=args.file.split(","))
|
importWp.fromFile(files=args.file.split(","))
|
||||||
if args.menu is False:
|
|
||||||
menuWp = WPMenu(name="Thread-1", basic=basic, wordpress=args.wordpress, logger=logger, parser=args.parser, ssl_canalblog=ssl_canalblog, ssl_wordpress=ssl_wordpress)
|
|
||||||
menuWp.fromFile("{0}".format(args.file.split(",")[0]))
|
|
||||||
if len(args.directory) > 0:
|
if len(args.directory) > 0:
|
||||||
|
remove(args, basic, logger)
|
||||||
try:
|
try:
|
||||||
with futures.ThreadPoolExecutor(max_workers=int(args.parallel)) as ex:
|
with futures.ThreadPoolExecutor(max_workers=int(args.parallel)) as ex:
|
||||||
wait_for = [ ex.submit(remove, i, args.parallel, args, basic, logger, ssl_wordpress) for i in range(0, int(args.parallel)) ]
|
|
||||||
wait(wait_for, return_when=ALL_COMPLETED)
|
|
||||||
errorRevert(logger, args.revert, args.tmp)
|
|
||||||
wait_for = [
|
wait_for = [
|
||||||
ex.submit(importDirectory, i, int(args.parallel), args.directory, logger, args.parser, args.wordpress, basic, args.serial, ssl_wordpress, args.create, args.update, args.image, args.revert, args.author)
|
ex.submit(importDirectory, i, int(args.parallel), args.directory, logger, args.parser, args.wordpress, basic, args.serial)
|
||||||
for i in range(0, int(args.parallel))
|
for i in range(0, int(args.parallel))
|
||||||
]
|
]
|
||||||
if args.menu is False:
|
|
||||||
wait(wait_for, return_when=ALL_COMPLETED)
|
|
||||||
menuWp = WPMenu(name="Thread-1", basic=basic, wordpress=args.wordpress, logger=logger, parser=args.parser, ssl_canalblog=ssl_canalblog, ssl_wordpress=ssl_wordpress)
|
|
||||||
menuWp.fromFile("{0}/index.html".format(args.directory))
|
|
||||||
except Exception as err:
|
except Exception as err:
|
||||||
logger.error("Threading error : {0}".format(err))
|
logger.error("Threading error : {0}".format(err))
|
||||||
if len(args.canalblog) > 0:
|
if len(args.canalblog) > 0:
|
||||||
|
remove(args, basic, logger)
|
||||||
try:
|
try:
|
||||||
with futures.ThreadPoolExecutor(max_workers=int(args.parallel)) as ex:
|
with futures.ThreadPoolExecutor(max_workers=int(args.parallel)) as ex:
|
||||||
wait_for = [ ex.submit(remove, i, args.parallel, args, basic, logger, ssl_wordpress) for i in range(0, int(args.parallel)) ]
|
|
||||||
wait(wait_for, return_when=ALL_COMPLETED)
|
|
||||||
errorRevert(logger, args.revert, args.tmp)
|
|
||||||
wait_for = [
|
wait_for = [
|
||||||
ex.submit(importUrl, i, int(args.parallel), args.canalblog, logger, args.parser, args.wordpress, basic, args.serial, ssl_wordpress, ssl_canalblog, args.create, args.update, args.image, args.revert, args.tmp, args.author)
|
ex.submit(importUrl, i, int(args.parallel), args.canalblog, logger, args.parser, args.wordpress, basic, args.serial)
|
||||||
for i in range(0, int(args.parallel))
|
for i in range(0, int(args.parallel))
|
||||||
]
|
]
|
||||||
if args.menu is False:
|
|
||||||
wait(wait_for, return_when=ALL_COMPLETED)
|
|
||||||
menuWp = WPMenu(name="Thread-1", basic=basic, wordpress=args.wordpress, logger=logger, parser=args.parser, ssl_canalblog=ssl_canalblog, ssl_wordpress=ssl_wordpress)
|
|
||||||
menuWp.fromUrl(args.canalblog)
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
except Exception as err:
|
except Exception as err:
|
||||||
logger.error("Threading error : {0}".format(err))
|
logger.error("Threading error : {0}".format(err))
|
||||||
exit(0)
|
exit(0)
|
||||||
@ -309,14 +234,11 @@ if __name__ == '__main__':
|
|||||||
|
|
||||||
if args.command == "export":
|
if args.command == "export":
|
||||||
canalblog = args.url.split(",")
|
canalblog = args.url.split(",")
|
||||||
protocol = "https"
|
exportWp = WPExport(logger=logger, parser=args.parser, directory=args.directory)
|
||||||
if ssl_canalblog is False:
|
|
||||||
protocol = "http"
|
|
||||||
exportWp = WPExport(logger=logger, parser=args.parser, directory=args.directory, ssl_canalblog=ssl_canalblog)
|
|
||||||
for canal in canalblog:
|
for canal in canalblog:
|
||||||
try:
|
try:
|
||||||
o = urlparse(canal)
|
o = urlparse(canal)
|
||||||
o = o._replace(scheme=protocol)
|
o = o._replace(scheme="https")
|
||||||
url = o.geturl().replace(":///", "://")
|
url = o.geturl().replace(":///", "://")
|
||||||
except Exception as err:
|
except Exception as err:
|
||||||
logger.error("parsing error : {0}".format(err))
|
logger.error("parsing error : {0}".format(err))
|
||||||
@ -335,7 +257,7 @@ if __name__ == '__main__':
|
|||||||
try:
|
try:
|
||||||
with futures.ThreadPoolExecutor(max_workers=int(args.parallel)) as ex:
|
with futures.ThreadPoolExecutor(max_workers=int(args.parallel)) as ex:
|
||||||
wait_for = [
|
wait_for = [
|
||||||
ex.submit(download, i, int(args.parallel), url, logger, args.parser, args.directory, args.html, args.img, ssl_canalblog, args.revert, args.tmp)
|
ex.submit(download, i, int(args.parallel), url, logger, args.parser, args.directory, args.html, args.img)
|
||||||
for i in range(0, int(args.parallel))
|
for i in range(0, int(args.parallel))
|
||||||
]
|
]
|
||||||
except Exception as err:
|
except Exception as err:
|
||||||
@ -344,31 +266,5 @@ if __name__ == '__main__':
|
|||||||
|
|
||||||
|
|
||||||
if args.command == "remove":
|
if args.command == "remove":
|
||||||
try:
|
remove(args, basic, logger)
|
||||||
with futures.ThreadPoolExecutor(max_workers=int(args.parallel)) as ex:
|
|
||||||
wait_for = [ ex.submit(remove, i, args.parallel, args, basic, logger, ssl_wordpress) for i in range(0, int(args.parallel)) ]
|
|
||||||
except Exception as err:
|
|
||||||
logger.error("Thread error for remove : {0}".format(err))
|
|
||||||
exit(0)
|
|
||||||
|
|
||||||
if args.command == "change":
|
|
||||||
if len(args.directory) > 0:
|
|
||||||
try:
|
|
||||||
with futures.ThreadPoolExecutor(max_workers=int(args.parallel)) as ex:
|
|
||||||
errorRevert(logger, args.revert, args.tmp)
|
|
||||||
wait_for = [ ex.submit(change, i, args.parallel, args, logger, args.tmp, args.revert) for i in range(0, int(args.parallel)) ]
|
|
||||||
except Exception as err:
|
|
||||||
logger.error("Thread error for remove : {0}".format(err))
|
|
||||||
if len(args.file) > 0:
|
|
||||||
changeWp = WPChange(logger=logger)
|
|
||||||
for filei in args.file.split(","):
|
|
||||||
changeWp.fromFile(filei)
|
|
||||||
exit(0)
|
|
||||||
|
|
||||||
if args.command == "menu":
|
|
||||||
menuWp = WPMenu(name="Thread-1", basic=basic, wordpress=args.wordpress, logger=logger, parser=args.parser, ssl_canalblog=ssl_canalblog, ssl_wordpress=ssl_wordpress)
|
|
||||||
if len(args.file) > 0:
|
|
||||||
menuWp.fromFile(args.file)
|
|
||||||
if len(args.canalblog) > 0:
|
|
||||||
menuWp.fromUrl(args.canalblog)
|
|
||||||
exit(0)
|
exit(0)
|
173
lib/WPChange.py
173
lib/WPChange.py
@ -1,173 +0,0 @@
|
|||||||
from bs4 import BeautifulSoup
|
|
||||||
from urllib.parse import urlparse
|
|
||||||
import requests, os, logging, re, json
|
|
||||||
|
|
||||||
class WPChange:
|
|
||||||
# Constructor
|
|
||||||
def __init__(self, index_name=1, number_thread=1, logger=None, parser="html.parser", tmp="/tmp/import_export_canablog"):
|
|
||||||
self._name = "Thread-{0}".format(index_name)
|
|
||||||
self._logger = logger
|
|
||||||
self._number_thread = number_thread
|
|
||||||
self._parser = parser
|
|
||||||
self._tmp = tmp
|
|
||||||
self._index_name = index_name
|
|
||||||
|
|
||||||
# Destructor
|
|
||||||
def __del__(self):
|
|
||||||
print("{0} : Import finished".format(self._name))
|
|
||||||
|
|
||||||
|
|
||||||
# Public method
|
|
||||||
|
|
||||||
## from file
|
|
||||||
|
|
||||||
def fromFile(self, files=[], number_thread=1, max_thread=1):
|
|
||||||
divFiles = int(len(files) / max_thread)
|
|
||||||
currentRangeFiles = int(divFiles * (number_thread))
|
|
||||||
firstRange = int(currentRangeFiles - divFiles)
|
|
||||||
self._logger.debug("{0} : index : {1}".format(self._name,number_thread))
|
|
||||||
|
|
||||||
|
|
||||||
self._logger.debug("{0} : first range : {1}".format(self._name,firstRange))
|
|
||||||
self._logger.debug("{0} : last range : {1}".format(self._name,currentRangeFiles))
|
|
||||||
|
|
||||||
for i in range(firstRange, currentRangeFiles):
|
|
||||||
|
|
||||||
if os.path.exists(files[i]):
|
|
||||||
self._logger.info("{0} : ({1}/{2}) File is being processed : {3}".format(self._name, i+1, currentRangeFiles + 1, files[i]))
|
|
||||||
self._change(files[i])
|
|
||||||
|
|
||||||
|
|
||||||
## From directory
|
|
||||||
|
|
||||||
|
|
||||||
def fromDirectory(self, directory="", revert=False):
|
|
||||||
self._directory = directory
|
|
||||||
directory = "{0}/archives".format(directory)
|
|
||||||
directories = self._getDirectories([], "{0}".format(directory))
|
|
||||||
if len(directories) > 0:
|
|
||||||
files = self._getFiles(directories)
|
|
||||||
if revert is False:
|
|
||||||
self._tmpFiles(files=files, number_thread=self._index_name, max_thread=self._number_thread)
|
|
||||||
self._fromFileTmp()
|
|
||||||
else:
|
|
||||||
self._logger.error("{0} : No files for {1}".format(self._name, directory))
|
|
||||||
|
|
||||||
|
|
||||||
def fromFile(self, files=[]):
|
|
||||||
for i in range(0, len(files)):
|
|
||||||
if os.path.exists(files[i]):
|
|
||||||
self._logger.info("{0} : ({1}/{2}) File is being processed : {3}".format(self._name, i+1, len(files), files[i]))
|
|
||||||
self._change(files[i])
|
|
||||||
|
|
||||||
|
|
||||||
# Private method
|
|
||||||
|
|
||||||
def _fromFileTmp(self):
|
|
||||||
try:
|
|
||||||
with open("{0}/{1}.json".format(self._tmp, self._name)) as file:
|
|
||||||
files = json.loads(file.read())
|
|
||||||
self._logger.debug("{0} : size of webpage : {1}".format(self._name, len(files)))
|
|
||||||
for i in range(0, len(files)):
|
|
||||||
if os.path.exists(files[i]):
|
|
||||||
self._logger.info("{0} : ({1}/{2}) File is being processed : {3}".format(self._name, i+1, len(files), files[i]))
|
|
||||||
self._change(files[i])
|
|
||||||
except Exception as ex:
|
|
||||||
self._logger.error("{0} : Read file json from tmp : {1}".format(self._name, ex))
|
|
||||||
|
|
||||||
|
|
||||||
def _tmpFiles(self, files=[], number_thread=1, max_thread=1):
|
|
||||||
print()
|
|
||||||
divFiles = int(len(files) / int(max_thread))
|
|
||||||
currentRangeFiles = int(divFiles * (int(number_thread)+1))
|
|
||||||
firstRange = int(currentRangeFiles - divFiles)
|
|
||||||
self._logger.debug("{0} : index : {1}".format(self._name,number_thread))
|
|
||||||
|
|
||||||
self._logger.debug("{0} : first range : {1}".format(self._name,firstRange))
|
|
||||||
self._logger.debug("{0} : last range : {1}".format(self._name,currentRangeFiles))
|
|
||||||
webpage = []
|
|
||||||
for i in range(firstRange, currentRangeFiles):
|
|
||||||
webpage.append(files[i])
|
|
||||||
|
|
||||||
try:
|
|
||||||
string_webpage = json.dumps(webpage)
|
|
||||||
open("{0}/{1}.json".format(self._tmp, self._name), "wt").write(string_webpage)
|
|
||||||
except Exception as ex:
|
|
||||||
self._logger.error("{0} : Error for writing webpage : {1}".format(self._name, ex))
|
|
||||||
|
|
||||||
## Get all files
|
|
||||||
|
|
||||||
def _getFiles(self, item):
|
|
||||||
files = []
|
|
||||||
for i in item:
|
|
||||||
for j in os.listdir(i):
|
|
||||||
if os.path.isfile("{0}/{1}".format(i, j)):
|
|
||||||
files.append("{0}/{1}".format(i, j))
|
|
||||||
return files
|
|
||||||
|
|
||||||
|
|
||||||
## Get directories
|
|
||||||
|
|
||||||
def _getDirectories(self, subdirectory, item):
|
|
||||||
sub = subdirectory
|
|
||||||
for i in os.listdir(item):
|
|
||||||
if os.path.isdir("{0}/{1}".format(item, i)):
|
|
||||||
sub.append("{0}/{1}".format(item, i))
|
|
||||||
subdirectory = self._getDirectories(sub, "{0}/{1}".format(item, i))
|
|
||||||
return subdirectory
|
|
||||||
|
|
||||||
## Change path img file
|
|
||||||
|
|
||||||
def _change(self, file):
|
|
||||||
ext_img = ["png", "svg", "gif", "jpg", "jpeg"]
|
|
||||||
try:
|
|
||||||
with open(file, 'r') as f:
|
|
||||||
content = f.read()
|
|
||||||
soup = BeautifulSoup(content, self._parser)
|
|
||||||
img = soup.find_all("img")
|
|
||||||
for i in img:
|
|
||||||
src = i.get("src")
|
|
||||||
o = urlparse(src)
|
|
||||||
if len(o.netloc) > 0:
|
|
||||||
self._logger.info("{0} : Change source image {1} /img/{2}/{3}".format(self._name, src, o.netloc, o.path))
|
|
||||||
content = content.replace(src, "/img/{0}/{1}".format(o.netloc, o.path))
|
|
||||||
script = soup.find_all("script", {"type": "text/javascript"})
|
|
||||||
for i in script:
|
|
||||||
src = i.get("src")
|
|
||||||
if src is not None:
|
|
||||||
o = urlparse(src)
|
|
||||||
if len(o.netloc) > 0:
|
|
||||||
self._logger.info("{0} : Change source js {1} /dists/js/{2}/{3}".format(self._name, src, o.netloc, o.path))
|
|
||||||
content = content.replace(src, "/dists/js/{0}/{1}".format(o.netloc, o.path))
|
|
||||||
link = soup.find_all("link", {"rel": "stylesheet"})
|
|
||||||
for i in link:
|
|
||||||
href = i.get("href")
|
|
||||||
if href is not None:
|
|
||||||
o = urlparse(href)
|
|
||||||
if len(o.netloc) > 0:
|
|
||||||
self._logger.info("{0} : Change source css {1} /dists/css/{2}/{3}".format(self._name, href, o.netloc, o.path))
|
|
||||||
content = content.replace(href, "/dists/css/{0}/{1}".format(o.netloc, o.path))
|
|
||||||
|
|
||||||
a = soup.find_all("a", {"target": "_blank"})
|
|
||||||
for i in a:
|
|
||||||
href = i.get("href")
|
|
||||||
if href is not None:
|
|
||||||
o = urlparse(href)
|
|
||||||
if len(o.netloc) > 0:
|
|
||||||
ext = o.path.split(".")[len(o.path.split("."))-1]
|
|
||||||
if ext in ext_img:
|
|
||||||
self._logger.info("{0} : Change a img {1} /img/{2}/{3}".format(self._name, href, o.netloc, o.path))
|
|
||||||
content = content.replace(href, "/img/{0}/{1}".format(o.netloc, o.path))
|
|
||||||
try:
|
|
||||||
with open(file, "w") as f:
|
|
||||||
self._logger.info("{0} : File write : {1}".format(self._name, file))
|
|
||||||
f.write(content)
|
|
||||||
except Exception as ex:
|
|
||||||
self._logger.error("{0} : Error for write file {1} : {2}".format(self._name, file, ex))
|
|
||||||
|
|
||||||
except Exception as ex:
|
|
||||||
self._logger.error("{0} : Error for read file {1} : {2}".format(self._name, file, ex))
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
340
lib/WPExport.py
340
lib/WPExport.py
@ -1,28 +1,26 @@
|
|||||||
#!/usr/bin/python3
|
#!/usr/bin/python3
|
||||||
from bs4 import BeautifulSoup
|
from bs4 import BeautifulSoup
|
||||||
from urllib.parse import urlparse
|
from urllib.parse import urlparse
|
||||||
import requests, os, argparse, logging, json
|
import requests, os, argparse, logging
|
||||||
from requests.adapters import HTTPAdapter
|
from requests.adapters import HTTPAdapter
|
||||||
from requests.packages.urllib3.util.retry import Retry
|
from requests.packages.urllib3.util.retry import Retry
|
||||||
|
|
||||||
class WPExport:
|
class WPExport:
|
||||||
def __init__(self, name = "Thread-0", url = "", logger = None, parser = "html.parser", directory = "backup", ssl_canalblog=True, tmp="/tmp/import_export_canablog"):
|
def __init__(self, name = "Thread-0", url = "", logger = None, parser = "html.parser", directory = "backup"):
|
||||||
self._url = url
|
self._url = url
|
||||||
self._logger = logger
|
self._logger = logger
|
||||||
self._parser = parser
|
self._parser = parser
|
||||||
self._dir = directory
|
self._dir = directory
|
||||||
self._name = name
|
self._name = name
|
||||||
self._protocol = "https"
|
|
||||||
if ssl_canalblog is False:
|
|
||||||
self._protocol = "http"
|
|
||||||
|
|
||||||
self._request = requests.Session()
|
self._request = requests.Session()
|
||||||
|
|
||||||
retries = Retry(total=10,
|
retries = Retry(total=10,
|
||||||
status_forcelist=[429, 500, 502, 503, 504], backoff_factor=2)
|
status_forcelist=[429, 500, 502, 503, 504], backoff_factor=2)
|
||||||
|
|
||||||
self._request.mount('{0}://'.format(self._protocol), HTTPAdapter(max_retries=retries))
|
self._request.mount('http://', HTTPAdapter(max_retries=retries))
|
||||||
self._tmp = tmp
|
|
||||||
|
|
||||||
# Destructor
|
# Destructor
|
||||||
def __del__(self):
|
def __del__(self):
|
||||||
@ -57,130 +55,104 @@ class WPExport:
|
|||||||
|
|
||||||
# Download HTML
|
# Download HTML
|
||||||
|
|
||||||
def downloadHTML(self, first, second):
|
def downloadHTML(self, webpage):
|
||||||
try:
|
self._downloadPage(webpage, self._dir)
|
||||||
with open("{0}/{1}.json".format(self._tmp, self._name)) as file:
|
|
||||||
webpage = json.loads(file.read())
|
|
||||||
self._downloadPage(webpage[first][second], self._dir)
|
|
||||||
except Exception as ex:
|
|
||||||
self._logger.error("{0} : Read file json from tmp : {1}".format(self._name, ex))
|
|
||||||
|
|
||||||
# Download Image
|
# Download Image
|
||||||
|
|
||||||
def downloadImg(self, first, second):
|
def downloadImg(self, webpage):
|
||||||
try:
|
page_src = self._getImg(webpage)
|
||||||
with open("{0}/{1}.json".format(self._tmp, self._name)) as file:
|
o = urlparse(self._url)
|
||||||
webpage = json.loads(file.read())
|
self._downloadPage(page_src, "{0}/{1}/{2}".format(self._dir, o.path, "img"))
|
||||||
page_src = self._getImg(webpage[first][second])
|
|
||||||
o = urlparse(self._url)
|
|
||||||
self._downloadPage(page_src, "{0}/{1}/{2}".format(self._dir, o.path, "img"))
|
|
||||||
except Exception as ex:
|
|
||||||
self._logger.error("{0} : Read file json from tmp : {1}".format(self._name, ex))
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
# Get URL
|
# Get URL
|
||||||
def getUrlPage(self, index_thread, max_thread):
|
def getUrlPage(self, index_thread, max_thread):
|
||||||
try:
|
try:
|
||||||
page = self._request.get(self._url)
|
page = self._request.get(self._url)
|
||||||
|
except Exception as err:
|
||||||
page_url = []
|
|
||||||
if page.status_code == 200:
|
|
||||||
soup = BeautifulSoup(page.text, self._parser)
|
|
||||||
ul = soup.find_all("ul", id="listsmooth")
|
|
||||||
for anchor in ul[0].find_all("a"):
|
|
||||||
href = anchor.get('href', '/')
|
|
||||||
if href != "#":
|
|
||||||
page_url.append(href)
|
|
||||||
else:
|
|
||||||
self._logger.error("{0} : URL did not get due status code : {1}".format(self._name, page.status_code))
|
|
||||||
self._logger.debug("{0} : {1}".format(self._name, page.content))
|
|
||||||
except ConnectionError as err:
|
|
||||||
self._logger.error("{0} : Connection error : {1}".format(self._name, err))
|
self._logger.error("{0} : Connection error : {1}".format(self._name, err))
|
||||||
exit(1)
|
exit(1)
|
||||||
except Exception as err:
|
page_url = []
|
||||||
self._logger.error("{0} : Exception error : {1}".format(self._name, err))
|
if page.status_code == 200:
|
||||||
|
soup = BeautifulSoup(page.text, self._parser)
|
||||||
|
ul = soup.find_all("ul", id="listsmooth")
|
||||||
|
for anchor in ul[0].find_all("a"):
|
||||||
|
href = anchor.get('href', '/')
|
||||||
|
if href != "#":
|
||||||
|
page_url.append(href)
|
||||||
|
else:
|
||||||
|
self._logger.error("{0} : URL did not get due status code : {1}".format(self._name, page.status_code))
|
||||||
|
self._logger.debug("{0} : {1}".format(self._name, page.content))
|
||||||
|
|
||||||
webpage = {"principal": {"page":[], "article":[]}, "publications": {"page":[], "article":[]}}
|
webpage = {"principal": {"page":[], "article":[]}, "publications": {"page":[], "article":[]}}
|
||||||
for i in page_url:
|
for i in page_url:
|
||||||
section = "publications"
|
section = "publications"
|
||||||
o = urlparse(i)
|
o = urlparse(i)
|
||||||
o = o._replace(scheme=self._protocol)
|
o = o._replace(scheme="https")
|
||||||
i = o.geturl().replace(":///", "://")
|
i = o.geturl().replace(":///", "://")
|
||||||
if i == "{0}/".format(self._url):
|
if i == "{0}/".format(self._url):
|
||||||
section = "principal"
|
section = "principal"
|
||||||
try:
|
try:
|
||||||
page = self._request.get(i)
|
page = self._request.get(i)
|
||||||
|
except Exception as err:
|
||||||
if page.status_code == 200:
|
|
||||||
self._logger.info("{0} : page : {1}".format(self._name, i))
|
|
||||||
if i not in webpage[section]["page"]:
|
|
||||||
webpage[section]["page"].append(i)
|
|
||||||
soup = BeautifulSoup(page.text, self._parser)
|
|
||||||
class_div = soup.find_all("div", class_="pagingfirstline")
|
|
||||||
if len(class_div) > 0:
|
|
||||||
pagingfirstline = class_div[0].find_all("a")
|
|
||||||
if len(pagingfirstline) > 1:
|
|
||||||
lastpage = pagingfirstline[len(pagingfirstline)-1].get("href", "/")
|
|
||||||
self._logger.debug("{0} : Last page {1}".format(self._name, lastpage))
|
|
||||||
|
|
||||||
element_lastpage = lastpage.split("/")[len(lastpage.split("/"))-1]
|
|
||||||
number_page = element_lastpage.split("-")[0].split("p")[1]
|
|
||||||
number_lastpage = int(number_page) / 10
|
|
||||||
|
|
||||||
setPageDivided = int(number_lastpage) / max_thread
|
|
||||||
if setPageDivided > int(setPageDivided):
|
|
||||||
setPageDivided = setPageDivided + 1
|
|
||||||
setPagePart = setPageDivided * (index_thread + 1)
|
|
||||||
firstPagePart = (setPagePart - setPageDivided)
|
|
||||||
|
|
||||||
self._logger.debug("{0} : Total page : {1}".format(self._name,int(number_lastpage)))
|
|
||||||
self._logger.debug("{0} : First range : {1}".format(self._name, int(firstPagePart)))
|
|
||||||
self._logger.debug("{0} : Last range : {1}".format(self._name, int(setPagePart)))
|
|
||||||
|
|
||||||
for j in range(int(firstPagePart),int(setPagePart)+1):
|
|
||||||
paging = j * 10
|
|
||||||
categorie = urlparse(i).path.split("/")
|
|
||||||
url_paging = "{0}/archives/p{1}-10.html".format(self._url, paging)
|
|
||||||
if len(categorie) > 2:
|
|
||||||
url_paging = "{0}/archives/{1}/p{2}-10.html".format(self._url, categorie[2], paging)
|
|
||||||
self._logger.info("{0} : {1}".format(self._name, url_paging))
|
|
||||||
if url_paging not in webpage[section]["page"]:
|
|
||||||
webpage[section]["page"].append(url_paging)
|
|
||||||
page = self._request.get(url_paging)
|
|
||||||
if page.status_code == 200:
|
|
||||||
soup = BeautifulSoup(page.text, self._parser)
|
|
||||||
h2 = soup.find_all("h2")
|
|
||||||
self._logger.debug("{0} : {1} H2 : {2}".format(self._name, url_paging, h2))
|
|
||||||
for title in h2:
|
|
||||||
self._logger.debug("{0} : {1} a : {2}".format(self._name, url_paging, title.find_all("a")))
|
|
||||||
href = title.find_all("a")[0].get("href", "/")
|
|
||||||
if href not in webpage[section]["article"]:
|
|
||||||
try:
|
|
||||||
o = urlparse(href)
|
|
||||||
o = o._replace(scheme="https").geturl()
|
|
||||||
webpage[section]["article"].append(o)
|
|
||||||
except Exception as err:
|
|
||||||
self._logger.error("parsing error : {0}".format(err))
|
|
||||||
exit(1)
|
|
||||||
else:
|
|
||||||
self._logger.error("{0} : web didn't get due status code : {1}".format(self._name, page.status_code))
|
|
||||||
self._logger.debug("{0} : {1}".format(self._name, page.content))
|
|
||||||
except ConnectionError as err:
|
|
||||||
self._logger.error("{0} : Connection error : {1}".format(self._name, err))
|
self._logger.error("{0} : Connection error : {1}".format(self._name, err))
|
||||||
exit(1)
|
exit(1)
|
||||||
except Exception as err:
|
if page.status_code == 200:
|
||||||
self._logger.error("{0} : Exception error : {1}".format(self._name, err))
|
self._logger.info("{0} : page : {1}".format(self._name, i))
|
||||||
exit(1)
|
if i not in webpage[section]["page"]:
|
||||||
try:
|
webpage[section]["page"].append(i)
|
||||||
string_webpage = json.dumps(webpage)
|
soup = BeautifulSoup(page.text, self._parser)
|
||||||
open("{0}/{1}.json".format(self._tmp, self._name), "wt").write(string_webpage)
|
class_div = soup.find_all("div", class_="pagingfirstline")
|
||||||
except Exception as ex:
|
if len(class_div) > 0:
|
||||||
self._logger.error("{0} : Error for writing webpage : {1}".format(self._name, ex))
|
pagingfirstline = class_div[0].find_all("a")
|
||||||
|
if len(pagingfirstline) > 1:
|
||||||
|
lastpage = pagingfirstline[len(pagingfirstline)-1].get("href", "/")
|
||||||
|
self._logger.debug("{0} : Last page {1}".format(self._name, lastpage))
|
||||||
|
|
||||||
|
element_lastpage = lastpage.split("/")[len(lastpage.split("/"))-1]
|
||||||
|
number_page = element_lastpage.split("-")[0].split("p")[1]
|
||||||
|
number_lastpage = int(number_page) / 10
|
||||||
|
|
||||||
|
setPageDivided = int(number_lastpage) / max_thread
|
||||||
|
if setPageDivided > int(setPageDivided):
|
||||||
|
setPageDivided = setPageDivided + 1
|
||||||
|
setPagePart = setPageDivided * (index_thread + 1)
|
||||||
|
firstPagePart = (setPagePart - setPageDivided)
|
||||||
|
|
||||||
|
self._logger.debug("{0} : Total page : {1}".format(self._name,int(number_lastpage)))
|
||||||
|
self._logger.debug("{0} : First range : {1}".format(self._name, int(firstPagePart)))
|
||||||
|
self._logger.debug("{0} : Last range : {1}".format(self._name, int(setPagePart)))
|
||||||
|
|
||||||
|
for j in range(int(firstPagePart),int(setPagePart)+1):
|
||||||
|
paging = j * 10
|
||||||
|
categorie = urlparse(i).path.split("/")
|
||||||
|
url_paging = "{0}/archives/p{1}-10.html".format(self._url, paging)
|
||||||
|
if len(categorie) > 2:
|
||||||
|
url_paging = "{0}/archives/{1}/p{2}-10.html".format(self._url, categorie[2], paging)
|
||||||
|
self._logger.info("{0} : {1}".format(self._name, url_paging))
|
||||||
|
if url_paging not in webpage[section]["page"]:
|
||||||
|
webpage[section]["page"].append(url_paging)
|
||||||
|
page = self._request.get(url_paging)
|
||||||
|
if page.status_code == 200:
|
||||||
|
soup = BeautifulSoup(page.text, self._parser)
|
||||||
|
h2 = soup.find_all("h2")
|
||||||
|
self._logger.debug("{0} : {1} H2 : {2}".format(self._name, url_paging, h2))
|
||||||
|
for title in h2:
|
||||||
|
self._logger.debug("{0} : {1} a : {2}".format(self._name, url_paging, title.find_all("a")))
|
||||||
|
href = title.find_all("a")[0].get("href", "/")
|
||||||
|
if href not in webpage[section]["article"]:
|
||||||
|
try:
|
||||||
|
o = urlparse(href)
|
||||||
|
o = o._replace(scheme="https").geturl()
|
||||||
|
except Exception as err:
|
||||||
|
self._logger.error("parsing error : {0}".format(err))
|
||||||
|
exit(1)
|
||||||
|
webpage[section]["article"].append(o)
|
||||||
|
else:
|
||||||
|
self._logger.error("{0} : web didn't get due status code : {1}".format(self._name, page.status_code))
|
||||||
|
self._logger.debug("{0} : {1}".format(self._name, page.content))
|
||||||
|
return webpage
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
@ -209,52 +181,47 @@ class WPExport:
|
|||||||
def _getScriptCss(self, js, css):
|
def _getScriptCss(self, js, css):
|
||||||
try:
|
try:
|
||||||
page = self._request.get(self._url)
|
page = self._request.get(self._url)
|
||||||
|
except Exception as err:
|
||||||
page_url = []
|
self._logger.error("Connection error : {0}".format(err))
|
||||||
if page.status_code == 200:
|
exit(1)
|
||||||
soup = BeautifulSoup(page.text, self._parser)
|
page_url = []
|
||||||
if js is True:
|
if page.status_code == 200:
|
||||||
script = soup.find_all("script")
|
soup = BeautifulSoup(page.text, self._parser)
|
||||||
for anchor in script:
|
if js is True:
|
||||||
src = anchor.get("src", "/")
|
script = soup.find_all("script")
|
||||||
if src != "/":
|
for anchor in script:
|
||||||
|
src = anchor.get("src", "/")
|
||||||
|
if src != "/":
|
||||||
|
try:
|
||||||
|
u = urlparse(self._url)
|
||||||
|
o = urlparse(src)
|
||||||
|
except Exception as err:
|
||||||
|
self._logger.error("parsing error : {0}".format(err))
|
||||||
|
exit(1)
|
||||||
|
if o.netloc == "":
|
||||||
|
o = o._replace(netloc=u.netloc)
|
||||||
|
o = o._replace(scheme=u.scheme)
|
||||||
|
page_url.append(o.geturl())
|
||||||
|
if css is True:
|
||||||
|
link = soup.find_all("link")
|
||||||
|
for anchor in link:
|
||||||
|
rel = anchor.get("rel")
|
||||||
|
if rel[0] == "stylesheet":
|
||||||
|
href = anchor.get("href", "/")
|
||||||
|
if href != "/":
|
||||||
try:
|
try:
|
||||||
u = urlparse(self._url)
|
u = urlparse(self._url)
|
||||||
o = urlparse(src)
|
o = urlparse(href)
|
||||||
if o.netloc == "":
|
|
||||||
o = o._replace(netloc=u.netloc)
|
|
||||||
o = o._replace(scheme=u.scheme)
|
|
||||||
page_url.append(o.geturl())
|
|
||||||
except Exception as err:
|
except Exception as err:
|
||||||
self._logger.error("parsing error : {0}".format(err))
|
self._logger.error("parsing error : {0}".format(err))
|
||||||
exit(1)
|
exit(1)
|
||||||
|
if o.netloc == "":
|
||||||
if css is True:
|
o = o._replace(netloc=u.netloc)
|
||||||
link = soup.find_all("link")
|
o = o._replace(scheme=u.scheme)
|
||||||
for anchor in link:
|
page_url.append(o.geturl())
|
||||||
rel = anchor.get("rel")
|
else:
|
||||||
if rel[0] == "stylesheet":
|
self._logger.error("JS or CSS did not get due status code : {0}".format(page.status_code))
|
||||||
href = anchor.get("href", "/")
|
self._logger.debug(page.content)
|
||||||
if href != "/":
|
|
||||||
try:
|
|
||||||
u = urlparse(self._url)
|
|
||||||
o = urlparse(href)
|
|
||||||
if o.netloc == "":
|
|
||||||
o = o._replace(netloc=u.netloc)
|
|
||||||
o = o._replace(scheme=u.scheme)
|
|
||||||
page_url.append(o.geturl())
|
|
||||||
except Exception as err:
|
|
||||||
self._logger.error("parsing error : {0}".format(err))
|
|
||||||
exit(1)
|
|
||||||
|
|
||||||
else:
|
|
||||||
self._logger.error("JS or CSS did not get due status code : {0}".format(page.status_code))
|
|
||||||
self._logger.debug(page.content)
|
|
||||||
except ConnectionError as err:
|
|
||||||
self._logger.error("Connection error : {0}".format(err))
|
|
||||||
exit(1)
|
|
||||||
except Exception as err:
|
|
||||||
self._logger.error("Exception error : {0}".format(err))
|
|
||||||
|
|
||||||
return page_url
|
return page_url
|
||||||
|
|
||||||
@ -265,25 +232,22 @@ class WPExport:
|
|||||||
for i in webpage:
|
for i in webpage:
|
||||||
try:
|
try:
|
||||||
page = self._request.get(i)
|
page = self._request.get(i)
|
||||||
|
except Exception as err:
|
||||||
if page.status_code == 200:
|
|
||||||
soup = BeautifulSoup(page.text, self._parser)
|
|
||||||
img = soup.find_all("img")
|
|
||||||
self._logger.info("{0} : image from page: {1} : ".format(self._name,i))
|
|
||||||
for anchor in img:
|
|
||||||
src = anchor.get("src", "/")
|
|
||||||
if src != "/":
|
|
||||||
if src not in page_img:
|
|
||||||
self._logger.info("{0} : image: {1} : ".format(self._name, src))
|
|
||||||
page_img.append(src)
|
|
||||||
else:
|
|
||||||
self._logger.error("{0} : Image did not get due status code : {1}".format(self._name, page.status_code))
|
|
||||||
self._logger.debug("{0} : {1}".format(self._name, page.content))
|
|
||||||
except ConnectionError as err:
|
|
||||||
self._logger.error("{0} : Connection error : {1}".format(self._name, err))
|
self._logger.error("{0} : Connection error : {1}".format(self._name, err))
|
||||||
exit(1)
|
exit(1)
|
||||||
except Exception as err:
|
if page.status_code == 200:
|
||||||
self._logger.error("{0} : Exception error : {1}".format(self._name, err))
|
soup = BeautifulSoup(page.text, self._parser)
|
||||||
|
img = soup.find_all("img")
|
||||||
|
self._logger.info("{0} : image from page: {1} : ".format(self._name,i))
|
||||||
|
for anchor in img:
|
||||||
|
src = anchor.get("src", "/")
|
||||||
|
if src != "/":
|
||||||
|
if src not in page_img:
|
||||||
|
self._logger.info("{0} : image: {1} : ".format(self._name, src))
|
||||||
|
page_img.append(src)
|
||||||
|
else:
|
||||||
|
self._logger.error("{0} : Image did not get due status code : {1}".format(self._name, page.status_code))
|
||||||
|
self._logger.debug("{0} : {1}".format(self._name, page.content))
|
||||||
|
|
||||||
return page_img
|
return page_img
|
||||||
|
|
||||||
@ -294,33 +258,29 @@ class WPExport:
|
|||||||
for i in range(0, len(webpage)):
|
for i in range(0, len(webpage)):
|
||||||
try:
|
try:
|
||||||
o = urlparse(webpage[i])
|
o = urlparse(webpage[i])
|
||||||
|
|
||||||
path_web = o.path.split("/")
|
|
||||||
filePageWeb = path_web[len(path_web)-1]
|
|
||||||
path_web.pop(len(path_web)-1)
|
|
||||||
dir_page_web = "/".join(path_web)
|
|
||||||
self._mkdirPath("{0}/{1}/{2}".format(backup_dir, o.netloc, dir_page_web))
|
|
||||||
try:
|
|
||||||
r = self._request.get(webpage[i])
|
|
||||||
|
|
||||||
if r.status_code == 200:
|
|
||||||
fileDownload = "{0}/{1}/index.html".format(backup_dir, o.netloc)
|
|
||||||
if len(dir_page_web) > 0 and len(filePageWeb) > 0:
|
|
||||||
fileDownload = "{0}/{1}{2}/{3}".format(backup_dir, o.netloc, dir_page_web, filePageWeb)
|
|
||||||
self._logger.info("{0} : {1}/{2} : {3}".format(self._name, i+1, len(webpage), fileDownload))
|
|
||||||
try:
|
|
||||||
open(fileDownload, "wb").write(r.content)
|
|
||||||
except Exception as err:
|
|
||||||
self._logger.error("file error : {0}".format(err))
|
|
||||||
exit(1)
|
|
||||||
else:
|
|
||||||
self._logger.error("Not download due status code : {0}".format(r.status_code))
|
|
||||||
self._logger.debug(r.content)
|
|
||||||
except ConnectionError as err:
|
|
||||||
self._logger.error("{0} : Connection error : {1}".format(self._name, err))
|
|
||||||
exit(1)
|
|
||||||
except Exception as err:
|
|
||||||
self._logger.error("{0} Exception error : {1}".format(self._name, err))
|
|
||||||
except Exception as err:
|
except Exception as err:
|
||||||
self._logger.error("parsing error : {0}".format(err))
|
self._logger.error("parsing error : {0}".format(err))
|
||||||
exit(1)
|
exit(1)
|
||||||
|
path_web = o.path.split("/")
|
||||||
|
filePageWeb = path_web[len(path_web)-1]
|
||||||
|
path_web.pop(len(path_web)-1)
|
||||||
|
dir_page_web = "/".join(path_web)
|
||||||
|
self._mkdirPath("{0}/{1}/{2}".format(backup_dir, o.netloc, dir_page_web))
|
||||||
|
try:
|
||||||
|
r = self._request.get(webpage[i])
|
||||||
|
except Exception as err:
|
||||||
|
self._logger.error("Connection error : {0}".format(err))
|
||||||
|
exit(1)
|
||||||
|
if r.status_code == 200:
|
||||||
|
fileDownload = "{0}/{1}/index.html".format(backup_dir, o.netloc)
|
||||||
|
if len(dir_page_web) > 0 and len(filePageWeb) > 0:
|
||||||
|
fileDownload = "{0}/{1}{2}/{3}".format(backup_dir, o.netloc, dir_page_web, filePageWeb)
|
||||||
|
self._logger.info("{0} : {1}/{2} : {3}".format(self._name, i+1, len(webpage), fileDownload))
|
||||||
|
try:
|
||||||
|
open(fileDownload, "wb").write(r.content)
|
||||||
|
except Exception as err:
|
||||||
|
self._logger.error("file error : {0}".format(err))
|
||||||
|
exit(1)
|
||||||
|
else:
|
||||||
|
self._logger.error("Not download due status code : {0}".format(r.status_code))
|
||||||
|
self._logger.debug(r.content)
|
||||||
|
1029
lib/WPImport.py
1029
lib/WPImport.py
File diff suppressed because it is too large
Load Diff
394
lib/WPMenu.py
394
lib/WPMenu.py
@ -1,394 +0,0 @@
|
|||||||
#!/usr/bin/python3
|
|
||||||
|
|
||||||
from bs4 import BeautifulSoup
|
|
||||||
from urllib.parse import urlparse
|
|
||||||
import requests, os, logging, re, json
|
|
||||||
from requests.adapters import HTTPAdapter
|
|
||||||
from requests.packages.urllib3.util.retry import Retry
|
|
||||||
|
|
||||||
class WPMenu:
|
|
||||||
# Constructor
|
|
||||||
def __init__(self, name="Thread-0", basic=None, canalblog="", wordpress="", logger=None, parser="html.parser", ssl_canalblog=True, ssl_wordpress=True):
|
|
||||||
self._name = name
|
|
||||||
self._basic = basic
|
|
||||||
self._canalblog = canalblog
|
|
||||||
self._wordpress = wordpress
|
|
||||||
self._logger = logger
|
|
||||||
self._parser = parser
|
|
||||||
self._headers_json = {'Content-Type': 'application/json; charset=utf-8', 'Accept':'application/json'}
|
|
||||||
self._protocol_wordpress = "https"
|
|
||||||
self._protocol_canalblog = "https"
|
|
||||||
self._directory = "backup"
|
|
||||||
|
|
||||||
if ssl_wordpress is False:
|
|
||||||
self._protocol_wordpress = "http"
|
|
||||||
if ssl_canalblog is False:
|
|
||||||
self._protocol_canalblog = "http"
|
|
||||||
self._request_canalblog = requests.Session()
|
|
||||||
self._request_wordpress = requests.Session()
|
|
||||||
|
|
||||||
retries = Retry(connect=10, read=10, redirect=5,
|
|
||||||
status_forcelist=[429, 500, 502, 503, 504], backoff_factor=2)
|
|
||||||
|
|
||||||
self._request_canalblog.mount('{0}://'.format(self._protocol_canalblog), HTTPAdapter(max_retries=retries))
|
|
||||||
self._request_wordpress.mount('{0}://'.format(self._protocol_wordpress), HTTPAdapter(max_retries=retries))
|
|
||||||
|
|
||||||
|
|
||||||
# Destructor
|
|
||||||
def __del__(self):
|
|
||||||
print("{0} : Import finished for {1}".format(self._name, self._wordpress))
|
|
||||||
|
|
||||||
# Public method
|
|
||||||
|
|
||||||
## From file
|
|
||||||
|
|
||||||
def fromFile(self, files):
|
|
||||||
if os.path.exists(files):
|
|
||||||
with open(files, 'r') as f:
|
|
||||||
self._logger.info("{0} : File is being processed : {1}".format(self._name, files))
|
|
||||||
content = f.read()
|
|
||||||
self._menu(content)
|
|
||||||
else:
|
|
||||||
self._logger.error("{0} : File isn't exist : {1}".format(self._name, files))
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
## Get from URL
|
|
||||||
|
|
||||||
def fromUrl(self, canalblog):
|
|
||||||
self._canalblog = canalblog
|
|
||||||
try:
|
|
||||||
o = urlparse(canalblog)
|
|
||||||
o = o._replace(scheme=self._protocol_canalblog)
|
|
||||||
i = o.geturl().replace(":///", "://")
|
|
||||||
page = self._request_canalblog.get(i)
|
|
||||||
if page.status_code == 200:
|
|
||||||
self._logger.info("{0} : Page web is being processed : {1}".format(self._name, i))
|
|
||||||
self._menu(page.content)
|
|
||||||
else:
|
|
||||||
self._logger.error("{0} : index didn't get due status code : {1}".format(self._name, page.status_code))
|
|
||||||
self._logger.debug("{0} : {1}".format(self._name, page.content))
|
|
||||||
except ConnectionError as err:
|
|
||||||
self._logger.error("{0} : Connection error for get url {1} : {2}".format(self._name, canalblog, err))
|
|
||||||
exit(1)
|
|
||||||
except Exception as err:
|
|
||||||
self._logger.error("{0} : Exception error for get url {1} : {2}".format(self._name, canalblog, err))
|
|
||||||
|
|
||||||
## replace caracter
|
|
||||||
|
|
||||||
def _replaceCaracter(self, title_rendered):
|
|
||||||
list_replace = {'’': "'", '–': '-', '…': '...', '« ': '"', ' »': '"', '« ': '"', ' »': '"', '’': "'", '"‘': "'"}
|
|
||||||
for old, new in list_replace.items():
|
|
||||||
title_rendered = title_rendered.replace(old, new)
|
|
||||||
return title_rendered
|
|
||||||
|
|
||||||
def _getIdfromTitlePost(self, content):
|
|
||||||
idMenu = {"id":0, "type":"", "link":""}
|
|
||||||
soup = BeautifulSoup(content, self._parser)
|
|
||||||
articletitle = soup.find_all("h2", class_="articletitle")
|
|
||||||
if len(articletitle) > 0:
|
|
||||||
articletitle = articletitle[0].get_text()
|
|
||||||
search = "posts"
|
|
||||||
post_type = "post"
|
|
||||||
if len(articletitle) == 0:
|
|
||||||
articletitle = soup.find_all("div", class_="albumbody")
|
|
||||||
if len(articletitle) > 0:
|
|
||||||
articletitle = articletitle[0].find("h2").get_text()
|
|
||||||
search = "pages"
|
|
||||||
post_type = "page"
|
|
||||||
|
|
||||||
exist = False
|
|
||||||
for index in range(1,10):
|
|
||||||
if exist is False:
|
|
||||||
params = {"search":articletitle, "per_page":100, "page":index}
|
|
||||||
try:
|
|
||||||
self._logger.debug("{0} : Get Url for {3} : {1} {2}".format(self._name, "{1}://{0}/wp-json/wp/v2/{2}".format(self._wordpress, self._protocol_wordpress, search), params, search))
|
|
||||||
page = self._request_wordpress.get("{1}://{0}/wp-json/wp/v2/{2}".format(self._wordpress, self._protocol_wordpress, search), auth=self._basic, params=params)
|
|
||||||
if page.status_code == 200:
|
|
||||||
result = page.json()
|
|
||||||
self._logger.info("{0} : Get content {2} : {1}".format(self._name, len(result), search))
|
|
||||||
if len(result) > 0:
|
|
||||||
for i in result:
|
|
||||||
title_rendered = i["title"]["rendered"]
|
|
||||||
if len(articletitle) != len(title_rendered):
|
|
||||||
title_rendered = self._replaceCaracter(title_rendered)
|
|
||||||
self._logger.debug("{0} : comparaison debug {1} {2}".format(self._name, articletitle, title_rendered))
|
|
||||||
if articletitle == title_rendered:
|
|
||||||
self._logger.debug("{0} : get {2} id : {1}".format(self._name, i, search))
|
|
||||||
idMenu = {"id":i["id"], "type":post_type, "link": i["link"]}
|
|
||||||
exist = True
|
|
||||||
else:
|
|
||||||
self._logger.debug("{0} : {2} {1}".format(self._name, result, len(result)))
|
|
||||||
break
|
|
||||||
elif page.status_code == 400:
|
|
||||||
self._logger.debug("{0} : {2} {1}".format(self._name, page.content, page.status_code))
|
|
||||||
break
|
|
||||||
else:
|
|
||||||
self._logger.error("{0} : Post didn't get due status code : {1}".format(self._name, page.status_code))
|
|
||||||
self._logger.debug("{0} : {1}".format(self._name, page.content))
|
|
||||||
except ConnectionError as err:
|
|
||||||
self._logger.error("{0} : Connection error for get content : {1}".format(self._name, err))
|
|
||||||
exit(1)
|
|
||||||
except Exception as err:
|
|
||||||
self._logger.error("{0} : Exception error for get content : {1} ".format(self._name, err))
|
|
||||||
return idMenu
|
|
||||||
|
|
||||||
def _getIdFromPost(self, href):
|
|
||||||
idMenu = {"id":0, "type":"", "link":""}
|
|
||||||
o = urlparse(href)
|
|
||||||
if len(o.netloc) > 0:
|
|
||||||
try:
|
|
||||||
page = self._request_canalblog.get(href)
|
|
||||||
if page.status_code == 200:
|
|
||||||
self._logger.info("{0} : Get content : {1}".format(self._name, href))
|
|
||||||
idMenu = self._getIdfromTitlePost(page.content)
|
|
||||||
else:
|
|
||||||
self._logger.error("{0} : {2} didn't get due status code : {1}".format(self._name, page.status_code, href))
|
|
||||||
self._logger.debug("{0} : {1}".format(self._name, page.content))
|
|
||||||
except ConnectionError as err:
|
|
||||||
self._logger.error("{0} : Connection error for get url {1} : {2}".format(self._name, href, err))
|
|
||||||
exit(1)
|
|
||||||
except Exception as err:
|
|
||||||
self._logger.error("{0} : Exception error for get url {1} : {2}".format(self._name, href, err))
|
|
||||||
else:
|
|
||||||
if os.path.exists("{0}/..{1}".format(self._directory, o.path)):
|
|
||||||
try:
|
|
||||||
content = open("{0}/..{1}".format(self._directory, o.path), "r")
|
|
||||||
idMenu = self._getIdfromTitlePost(content)
|
|
||||||
except Exception as err:
|
|
||||||
self._logger.error("{0} : Exception error for get file content {1} : {2}".format(self._name, href, err))
|
|
||||||
return idMenu
|
|
||||||
|
|
||||||
|
|
||||||
def _getIdFromReverse(self, title, href):
|
|
||||||
self._logger.info("{0} : get title {1} from href {2}".format(self._name, title, href))
|
|
||||||
idMenu = {"id":0, "type":"", "link":""}
|
|
||||||
if href != "#":
|
|
||||||
title = href[::-1]
|
|
||||||
second_title = title.split("/")[2]
|
|
||||||
second_title = second_title[::-1]
|
|
||||||
link = title.split("/")[0]
|
|
||||||
link = link[::-1]
|
|
||||||
title = title.split("/")[1]
|
|
||||||
title = title[::-1]
|
|
||||||
self._logger.info("{0} link {1} title {2}".format(self._name, link, title))
|
|
||||||
if link == "index.html":
|
|
||||||
if second_title == "albums":
|
|
||||||
idMenu = self._getIdFromPost(href)
|
|
||||||
else:
|
|
||||||
idMenu = self._getId(title)
|
|
||||||
else:
|
|
||||||
idMenu = self._getIdFromPost(href)
|
|
||||||
|
|
||||||
return idMenu
|
|
||||||
|
|
||||||
def _getId(self, title):
|
|
||||||
idMenu = {"id": 0, "type":"", "link":""}
|
|
||||||
exist = False
|
|
||||||
if exist is False:
|
|
||||||
for i in ["categories", "tags"]:
|
|
||||||
typeId = "category"
|
|
||||||
if i == "tags":
|
|
||||||
typeId = "tag"
|
|
||||||
for index in range(1,10):
|
|
||||||
try:
|
|
||||||
params = {"search":title, "per_page":"100", "page":index}
|
|
||||||
self._logger.info("{0} Get menu {1} {2} {3}".format(self._name, "{2}://{0}/wp-json/wp/v2/{1}".format(self._wordpress, i, self._protocol_wordpress), index, title))
|
|
||||||
page = self._request_wordpress.get("{2}://{0}/wp-json/wp/v2/{1}".format(self._wordpress, i, self._protocol_wordpress), auth=self._basic, params=params)
|
|
||||||
if page.status_code == 200:
|
|
||||||
result = page.json()
|
|
||||||
if len(result) > 0:
|
|
||||||
for j in result:
|
|
||||||
self._logger.info("{0} info : {1} {2} {3}".format(self._name, j["name"], j["slug"], title))
|
|
||||||
if j["name"] == title or j["slug"] == title:
|
|
||||||
self._logger.info("{0} : comparaison ok : {1} {2}".format(self._name, j["id"], i))
|
|
||||||
|
|
||||||
idMenu = {"id": j["id"], "type": typeId, "link": j["link"]}
|
|
||||||
exist = True
|
|
||||||
else:
|
|
||||||
break
|
|
||||||
elif page.status_code == 400:
|
|
||||||
break
|
|
||||||
else:
|
|
||||||
self._logger.error("{0} : {2} didn't get due status code : {1}".format(self._name, page.status_code, i))
|
|
||||||
self._logger.debug("{0} : {1}".format(self._name, page.content))
|
|
||||||
except ConnectionError as err:
|
|
||||||
self._logger.error("{0} : Connection error for get url {1} : {2}".format(self._name, "{2}://{0}/wp-json/wp/v2/{1}".format(self._wordpress, i, self._protocol_wordpress), err))
|
|
||||||
exit(1)
|
|
||||||
except Exception as err:
|
|
||||||
self._logger.error("{0} : Exception error for get url {1} : {2}".format(self._name, "{2}://{0}/wp-json/wp/v2/{1}".format(self._wordpress, i, self._protocol_wordpress), err))
|
|
||||||
return idMenu
|
|
||||||
|
|
||||||
def _menu(self, content):
|
|
||||||
|
|
||||||
soup = BeautifulSoup(content, self._parser)
|
|
||||||
ul = soup.find("ul", id="listsmooth")
|
|
||||||
menu = list()
|
|
||||||
children = list()
|
|
||||||
for anchor in ul.find_all("li"):
|
|
||||||
|
|
||||||
parent = anchor.find("a").get_text().replace(" \xa0", "")
|
|
||||||
href = anchor.find("a").get("href")
|
|
||||||
if href == "{0}://{1}/".format(self._protocol_canalblog, self._canalblog):
|
|
||||||
parent = "home"
|
|
||||||
itemMenu = {"id":"", "type":"", "title": parent, "link":"", "href":href, "children":list()}
|
|
||||||
if href == "#":
|
|
||||||
li = anchor.find("ul").find_all("li")
|
|
||||||
for child in li:
|
|
||||||
a = child.find("a")
|
|
||||||
href = a.get("href")
|
|
||||||
self._logger.info("{0} Parent {1} : Child {2}".format(self._name, parent, a.get_text()))
|
|
||||||
children.append({"title": a.get_text(), "parent": parent, "href":href, "link":""})
|
|
||||||
menu.append(itemMenu)
|
|
||||||
|
|
||||||
|
|
||||||
for i in range(0, len(children)):
|
|
||||||
self._logger.info("{0} : Child {1} {2}".format(self._name, children[i], i))
|
|
||||||
for j in range(0, len(menu)):
|
|
||||||
if j < len(menu):
|
|
||||||
if menu[j]["title"] == children[i]["title"]:
|
|
||||||
self._logger.info("{0} : Parent {1} {2}".format(self._name, menu[j], j))
|
|
||||||
del menu[j]
|
|
||||||
|
|
||||||
for j in range(0, len(menu)):
|
|
||||||
self._logger.info("{0} : Children for : {1}".format(self._name, menu[j]["title"]))
|
|
||||||
if menu[j]["title"] == children[i]["parent"]:
|
|
||||||
menu[j]["children"].append({"id":"", "type":"", "title":children[i]["title"], "parent": children[i]["parent"], "link":"", "href":children[i]["href"]})
|
|
||||||
|
|
||||||
for i in range(0, len(menu)):
|
|
||||||
self._logger.info("{0} : Menu {1} {2}".format(self._name, menu[i]["title"], len(menu[i]["children"])))
|
|
||||||
if menu[i]["title"] != "home":
|
|
||||||
for j in range(0, len(menu[i]["children"])):
|
|
||||||
idMenu = self._getId(menu[i]["children"][j]["title"])
|
|
||||||
if idMenu["id"] == 0:
|
|
||||||
self._logger.debug("{0} : content children {1}".format(self._name, menu[i]["children"][j]))
|
|
||||||
idMenu = self._getIdFromReverse(menu[i]["children"][j]["title"], menu[i]["children"][j]["href"])
|
|
||||||
if idMenu["id"] != 0:
|
|
||||||
menu[i]["children"][j] = {"id":idMenu["id"], "type": idMenu["type"], "link": idMenu["link"], "title": menu[i]["children"][j]["title"], "parent": menu[i]["children"][j]["parent"]}
|
|
||||||
idMenu = self._getId(menu[i]["title"])
|
|
||||||
self._logger.debug("{0} : content parent {1}".format(self._name, menu[i]))
|
|
||||||
self._logger.debug("{0} : content idMenu {1}".format(self._name, idMenu))
|
|
||||||
|
|
||||||
if idMenu["id"] == 0:
|
|
||||||
idMenu = self._getIdFromReverse(menu[i]["title"], menu[i]["href"])
|
|
||||||
if idMenu["id"] != 0:
|
|
||||||
menu[i] = {"id":idMenu["id"], "type": idMenu["type"], "title":menu[i]["title"], "link":idMenu["link"], "children": menu[i]["children"]}
|
|
||||||
|
|
||||||
self._createMenu(menu)
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
def _createItemMenu(self, idMenu, itemMenu, parent):
|
|
||||||
idItemMenu = 0
|
|
||||||
self._logger.info("{0} : Create item menu from API Wordpress : {1}".format(self._name, self._wordpress))
|
|
||||||
try:
|
|
||||||
params = {"search": itemMenu["title"], "menus": idMenu}
|
|
||||||
page = self._request_wordpress.get("{1}://{0}/wp-json/wp/v2/menu-items".format(self._wordpress, self._protocol_wordpress), auth=self._basic, params=params)
|
|
||||||
if page.status_code == 200:
|
|
||||||
result = page.json()
|
|
||||||
for i in result:
|
|
||||||
if self._replaceCaracter(i["title"]["rendered"]) == itemMenu["title"]:
|
|
||||||
idItemMenu = int(i["id"])
|
|
||||||
self._logger.info("{0} : Length of result for item menus : {1}".format(self._name, len(result)))
|
|
||||||
url = "{1}://{0}/wp-json/wp/v2/menu-items".format(self._wordpress, self._protocol_wordpress)
|
|
||||||
if idItemMenu != 0:
|
|
||||||
url = "{1}://{0}/wp-json/wp/v2/menu-items/{2}".format(self._wordpress, self._protocol_wordpress, idItemMenu)
|
|
||||||
|
|
||||||
try:
|
|
||||||
objectt = itemMenu["type"]
|
|
||||||
if objectt == "tag":
|
|
||||||
objectt = "post_tag"
|
|
||||||
data = {"title": itemMenu["title"], "status": "publish", "parent":parent, "menus":idMenu, "url":"#"}
|
|
||||||
if itemMenu["title"] == "home":
|
|
||||||
data = {"title": itemMenu["title"], "status": "publish", "parent":parent, "menus":idMenu, "url":"{0}://{1}".format(self._protocol_wordpress, self._wordpress)}
|
|
||||||
if type(itemMenu["id"]) is str:
|
|
||||||
if len(itemMenu["id"]) > 0:
|
|
||||||
data = {"title": itemMenu["title"], "status": "publish", "parent":parent, "url": itemMenu["link"], "menus":idMenu, "object":objectt, "object_id":int(itemMenu["id"])}
|
|
||||||
elif type(itemMenu["id"]) is int:
|
|
||||||
data = {"title": itemMenu["title"], "status": "publish", "parent":parent, "url": itemMenu["link"], "menus":idMenu, "object":objectt, "object_id":itemMenu["id"]}
|
|
||||||
self._logger.debug("{0} : data for create/update : {1}".format(self._name, data))
|
|
||||||
page = self._request_wordpress.post(url, auth=self._basic, headers=self._headers_json, data=json.dumps(data))
|
|
||||||
|
|
||||||
if page.status_code in [201, 200]:
|
|
||||||
result = page.json()
|
|
||||||
idItemMenu = int(result["id"])
|
|
||||||
self._logger.info("{0} : create/update item menu : {1}".format(self._name, itemMenu["title"]))
|
|
||||||
else:
|
|
||||||
self._logger.error("{0} : Create menu items for {2} didn't get due status code : {1}".format(self._name, page.status_code, itemMenu["title"]))
|
|
||||||
self._logger.debug("{0} : {1} {2}".format(self._name, page.content, itemMenu))
|
|
||||||
|
|
||||||
except ConnectionError as err:
|
|
||||||
self._logger.error("{0} : Connection error for create item menu {1} : {2}".format(self._name, "{1}://{0}/wp-json/wp/v2/menu-items".format(self._wordpress, self._protocol_wordpress), err))
|
|
||||||
exit(1)
|
|
||||||
except Exception as err:
|
|
||||||
self._logger.error("{0} : Exception error for create item menu {1} : {2}".format(self._name, "{1}://{0}/wp-json/wp/v2/menu-items".format(self._wordpress, self._protocol_wordpress), err))
|
|
||||||
else:
|
|
||||||
self._logger.error("{0} : Get menu items didn't get due status code : {1}".format(self._name, page.status_code))
|
|
||||||
self._logger.debug("{0} : {1}".format(self._name, page.content))
|
|
||||||
except ConnectionError as err:
|
|
||||||
self._logger.error("{0} : Connection error for get item menus {1} : {2}".format(self._name, "{1}://{0}/wp-json/wp/v2/menu-items".format(self._wordpress, self._protocol_wordpress), err))
|
|
||||||
exit(1)
|
|
||||||
except Exception as err:
|
|
||||||
self._logger.error("{0} : Exception error for get item menus {1} : {2}".format(self._name, "{1}://{0}/wp-json/wp/v2/menu-items".format(self._wordpress, self._protocol_wordpress), err))
|
|
||||||
return idItemMenu
|
|
||||||
|
|
||||||
def _createMenu(self, menu):
|
|
||||||
title = "Menu {0}".format(self._wordpress)
|
|
||||||
self._logger.info("{0} : Create menu from API Wordpress : {1}".format(self._name, title))
|
|
||||||
try:
|
|
||||||
params = {"search": title}
|
|
||||||
page = self._request_wordpress.get("{1}://{0}/wp-json/wp/v2/menus".format(self._wordpress, self._protocol_wordpress), auth=self._basic, params=params)
|
|
||||||
if page.status_code == 200:
|
|
||||||
result = page.json()
|
|
||||||
self._logger.info("{0} : Get content menus : {1}".format(self._name, len(result)))
|
|
||||||
idMenu = 0
|
|
||||||
if len(result) == 0:
|
|
||||||
|
|
||||||
self._logger.info("{0} : Create menu : {1}".format(self._name, title))
|
|
||||||
data = {"name": title}
|
|
||||||
try:
|
|
||||||
page = self._request_wordpress.post("{1}://{0}/wp-json/wp/v2/menus".format(self._wordpress, self._protocol_wordpress), auth=self._basic, headers=self._headers_json, data=json.dumps(data))
|
|
||||||
if page.status_code == 201:
|
|
||||||
result = page.json()
|
|
||||||
self._logger.debug("{0} : Get menus : {1}".format(self._name, result))
|
|
||||||
|
|
||||||
if len(result) > 0:
|
|
||||||
idMenu = result["id"]
|
|
||||||
else:
|
|
||||||
self._logger.error("{0} : Post menu didn't get due status code : {1}".format(self._name, page.status_code))
|
|
||||||
self._logger.debug("{0} : {1}".format(self._name, page.content))
|
|
||||||
|
|
||||||
except ConnectionError as err:
|
|
||||||
self._logger.error("{0} : Connection error for create menu {1} : {2}".format(self._name, "{1}://{0}/wp-json/wp/v2/menus".format(self._wordpress, self._protocol_wordpress), err))
|
|
||||||
exit(1)
|
|
||||||
except Exception as err:
|
|
||||||
self._logger.error("{0} : Exception error for get menu {1} : {2}".format(self._name, "{1}://{0}/wp-json/wp/v2/menus".format(self._wordpress, self._protocol_wordpress), err))
|
|
||||||
else:
|
|
||||||
self._logger.debug("{0} : Get menus : {1}".format(self._name, result))
|
|
||||||
for i in result:
|
|
||||||
self._logger.debug("{0} : List menus : {1}".format(self._name, i))
|
|
||||||
if i["name"] == title:
|
|
||||||
idMenu = i["id"]
|
|
||||||
self._logger.info("{0} : Get ID menus : {1}".format(self._name, idMenu))
|
|
||||||
self._addItemMenu(menu, idMenu)
|
|
||||||
else:
|
|
||||||
self._logger.error("{0} : Get menu didn't get due status code : {1}".format(self._name, page.status_code))
|
|
||||||
self._logger.debug("{0} : {1}".format(self._name, page.content))
|
|
||||||
|
|
||||||
except ConnectionError as err:
|
|
||||||
self._logger.error("{0} : Connection error for get menu {1} : {2}".format(self._name, "{1}://{0}/wp-json/wp/v2/menus".format(self._wordpress, self._protocol_wordpress), err))
|
|
||||||
exit(1)
|
|
||||||
except Exception as err:
|
|
||||||
self._logger.error("{0} : Exception error for get menu {1} : {2}".format(self._name, "{1}://{0}/wp-json/wp/v2/menus".format(self._wordpress, self._protocol_wordpress), err))
|
|
||||||
|
|
||||||
|
|
||||||
def _addItemMenu(self, menu, idMenu):
|
|
||||||
self._logger.info("{0} : add item to menu : {1}".format(self._name, idMenu))
|
|
||||||
parent = 0
|
|
||||||
for i in menu:
|
|
||||||
parent = 0
|
|
||||||
self._logger.debug("{0} : debug create item menu : {1}".format(self._name, i))
|
|
||||||
parent = self._createItemMenu(idMenu, i, parent)
|
|
||||||
for j in i["children"]:
|
|
||||||
self._createItemMenu(idMenu, j, parent)
|
|
111
lib/WPRemove.py
111
lib/WPRemove.py
@ -8,23 +8,19 @@ from requests.packages.urllib3.util.retry import Retry
|
|||||||
|
|
||||||
class WPRemove:
|
class WPRemove:
|
||||||
# Constructor
|
# Constructor
|
||||||
def __init__(self, index_name=1, number_thread=1, basic=None, wordpress="", logger=None, ssl_wordpress=True):
|
def __init__(self, name="Thread-0", basic=None, wordpress="", logger=None):
|
||||||
|
self._name = name
|
||||||
self._basic = basic
|
self._basic = basic
|
||||||
self._wordpress = wordpress
|
self._wordpress = wordpress
|
||||||
self._logger = logger
|
self._logger = logger
|
||||||
self._headers_json = {'Content-Type': 'application/json', 'Accept':'application/json'}
|
self._headers_json = {'Content-Type': 'application/json', 'Accept':'application/json'}
|
||||||
self._name = "Thread-{0}".format(index_name)
|
|
||||||
self._index_thread = index_name
|
|
||||||
self._protocol = "https"
|
|
||||||
self._number_thread = number_thread
|
|
||||||
if ssl_wordpress is False:
|
|
||||||
self._protocol = "http"
|
|
||||||
self._request = requests.Session()
|
self._request = requests.Session()
|
||||||
|
|
||||||
retries = Retry(connect=10, read=10, redirect=5,
|
retries = Retry(connect=10, read=10, redirect=5,
|
||||||
status_forcelist=[429, 500, 502, 503, 504], backoff_factor=2)
|
status_forcelist=[429, 500, 502, 503, 504], backoff_factor=2)
|
||||||
|
|
||||||
self._request.mount('{0}://'.format(self._protocol), HTTPAdapter(max_retries=retries))
|
self._request.mount('http://', HTTPAdapter(max_retries=retries))
|
||||||
|
|
||||||
|
|
||||||
# Destructor
|
# Destructor
|
||||||
@ -33,24 +29,6 @@ class WPRemove:
|
|||||||
|
|
||||||
# Public method
|
# Public method
|
||||||
|
|
||||||
def _getCount(self, composant):
|
|
||||||
count = 0
|
|
||||||
try:
|
|
||||||
params = {"per_page":1}
|
|
||||||
self._logger.info("{0} : Get count {2} to remove for url : {1}".format(self._name, self._wordpress, composant))
|
|
||||||
r = self._request.get("{2}://{0}/wp-json/wp/v2/{1}".format(self._wordpress, composant, self._protocol), params=params, auth=self._basic, headers=self._headers_json)
|
|
||||||
if r.status_code == 200:
|
|
||||||
count = int(r.headers["X-WP-Total"])
|
|
||||||
else:
|
|
||||||
self._logger.error("{0} : Error for list to remove {1} due status code {2}".format(self._name, composant, r.status_code))
|
|
||||||
self._logger.debug("{0} : Content error for {1} : {2}".format(self._name, composant, r.content))
|
|
||||||
except ConnectionError as err:
|
|
||||||
self._logger.error("{0} : Connection error for list {1} to remove : {2}".format(self._name, composant, err))
|
|
||||||
exit(1)
|
|
||||||
except Exception as err:
|
|
||||||
self._logger.error("{0} : Exception error for list {1} to remove : {2}".format(self._name, composant, err))
|
|
||||||
return count
|
|
||||||
|
|
||||||
def setUrl(self, wordpress):
|
def setUrl(self, wordpress):
|
||||||
self._wordpress = wordpress
|
self._wordpress = wordpress
|
||||||
|
|
||||||
@ -69,60 +47,29 @@ class WPRemove:
|
|||||||
# Private method
|
# Private method
|
||||||
|
|
||||||
def _removeAll(self, composant):
|
def _removeAll(self, composant):
|
||||||
count = self._getCount(composant)
|
params = {"per_page":100}
|
||||||
self._logger.debug("{0} : Count for {1} : {2}".format(self._name, composant, count))
|
try:
|
||||||
if count > 0:
|
self._logger.info("{0} : List {2} to remove for url : {1}".format(self._name, self._wordpress, composant))
|
||||||
self._logger.debug("{0} : Number thread for {1} : {2}".format(self._name, composant, self._number_thread))
|
|
||||||
|
|
||||||
page = count / int(self._number_thread)
|
r = self._request.get("http://{0}/wp-json/wp/v2/{1}".format(self._wordpress, composant), auth=self._basic, params=params, headers=self._headers_json)
|
||||||
self._logger.debug("{0} : Page for {1} : {2}".format(self._name, composant, page))
|
except Exception as err:
|
||||||
|
self._logger.error("{0} : Connection error for list {1} to remove : {2}".format(self._name, composant, err))
|
||||||
if page > int(page):
|
if r.status_code == 200:
|
||||||
page = int(page) + 1
|
result = r.json()
|
||||||
if page > 100:
|
if len(result) > 0:
|
||||||
page = 100
|
for i in result:
|
||||||
params = {"per_page":page, "page":self._index_thread}
|
self._logger.info("{0} : Remove {2} for url {1} : {3}".format(self._name, self._wordpress, composant, i["title"]["rendered"]))
|
||||||
self._logger.info("{0} : Params for {1} : {2}".format(self._name, composant, params))
|
params = {"force":1}
|
||||||
|
try:
|
||||||
try:
|
r = self._request.delete("http://{0}/wp-json/wp/v2/{1}/{2}".format(self._wordpress, composant, i["id"]), auth=self._basic, headers=self._headers_json , params=params)
|
||||||
self._logger.info("{0} : List {2} to remove for url : {1}".format(self._name, self._wordpress, composant))
|
if r.status_code == 200:
|
||||||
|
self._logger.info("{0} : Post removed for URL {1} {2} : {3}".format(self._name, self._wordpress, composant, i["title"]["rendered"]))
|
||||||
r = self._request.get("{2}://{0}/wp-json/wp/v2/{1}".format(self._wordpress, composant, self._protocol), auth=self._basic, params=params, headers=self._headers_json)
|
else:
|
||||||
|
self._logger.error("{0} : Connection error for post {1} {2} {3} with status code {4}".format(self._name, self._wordpress, composant, i["title"]["rendered"], r.status_code))
|
||||||
if r.status_code == 200:
|
except Exception as err:
|
||||||
result = r.json()
|
self._logger.error("{0} : Connection error for {1} remove : {2}".format(self._name, composant, err))
|
||||||
if len(result) > 0:
|
exit(1)
|
||||||
for i in result:
|
self._removeAll(composant)
|
||||||
is_delete = True
|
else:
|
||||||
self._logger.info(i["slug"])
|
self._logger.error("{0} : Error for list to remove {1} due status code {2}".format(self._name, composant, r.status_code))
|
||||||
if i["slug"] == "non-classe":
|
self._logger.debug("{0} : Content error for {1} : {2}".format(self._name, composant, r.content))
|
||||||
is_delete = False
|
|
||||||
if is_delete is True:
|
|
||||||
if composant == "tags" or composant == "categories":
|
|
||||||
title = i["name"]
|
|
||||||
else:
|
|
||||||
title = i["title"]["rendered"]
|
|
||||||
self._logger.info("{0} : Remove {2} for url {1} : {3}".format(self._name, self._wordpress, composant, title))
|
|
||||||
params = {"force":1}
|
|
||||||
try:
|
|
||||||
r = self._request.delete("{3}://{0}/wp-json/wp/v2/{1}/{2}".format(self._wordpress, composant, i["id"], self._protocol), auth=self._basic, headers=self._headers_json , params=params)
|
|
||||||
if r.status_code == 200:
|
|
||||||
self._logger.info("{0} : Post removed for URL {1} {2} : {3}".format(self._name, self._wordpress, composant, title))
|
|
||||||
else:
|
|
||||||
self._logger.error("{0} : Connection error for post {1} {2} {3} with status code {4}".format(self._name, self._wordpress, composant, title, r.status_code))
|
|
||||||
except ConnectionError as err:
|
|
||||||
self._logger.error("{0} : Connection error for {1} remove : {2}".format(self._name, composant, err))
|
|
||||||
exit(1)
|
|
||||||
except Exception as err:
|
|
||||||
self._logger.error("{0} : Exception error for {1} remove : {2}".format(self._name, composant, err))
|
|
||||||
self._removeAll(composant)
|
|
||||||
if r.status_code == 400:
|
|
||||||
self._logger.error("{0} : No content for {1} to remove : {2}".format(self._name, composant, r.status_code))
|
|
||||||
else:
|
|
||||||
self._logger.error("{0} : Error for list to remove {1} due status code {2}".format(self._name, composant, r.status_code))
|
|
||||||
self._logger.debug("{0} : Content error for {1} : {2}".format(self._name, composant, r.content))
|
|
||||||
except ConnectionError as err:
|
|
||||||
self._logger.error("{0} : Connection error for list {1} to remove : {2}".format(self._name, composant, err))
|
|
||||||
exit(1)
|
|
||||||
except Exception as err:
|
|
||||||
self._logger.error("{0} : Exception error for list {1} to remove : {2}".format(self._name, composant, err))
|
|
||||||
|
Loading…
x
Reference in New Issue
Block a user