Skip to content

Commit 03e9902

Browse files
committed
修复import错误,修复EXCEL工具类必须要文件目录的异常
1 parent dfa041f commit 03e9902

File tree

9 files changed

+26
-16
lines changed

9 files changed

+26
-16
lines changed

src/PaperCrawlerUtil/common_util.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22
import sys
33
from typing import List
44

5-
import global_val
5+
import PaperCrawlerUtil.global_val
66
from proxypool.storages.proxy_dict import ProxyDict
77

88
sys.path.append("PaperCrawlerUtil")

src/PaperCrawlerUtil/office_util.py

Lines changed: 18 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -98,7 +98,7 @@ class ExcelProcess:
9898

9999
#@todo 只能打开xls,需要换库解决,xlrd新版本不行了
100100

101-
def __init__(self, filename: str, sheet_index: int = 0, sheet_name: str = None) -> None:
101+
def __init__(self, filename: str = "", sheet_index: int = 0, sheet_name: str = None) -> None:
102102

103103
"""
104104
只能处理xls文件,后续进行改造
@@ -110,20 +110,30 @@ def __init__(self, filename: str, sheet_index: int = 0, sheet_name: str = None)
110110
self.filename = filename
111111
self.sheet_index = sheet_index
112112
self.sheet_name = sheet_name
113-
self.excel = xlrd.open_workbook(filename)
114-
self.sheet_names = self.excel.sheet_names()
115-
self.sheet = self.excel.sheet_by_name(sheet_name) \
116-
if sheet_name is not None else self.excel.sheet_by_index(sheet_index)
117-
self.row_size = self.sheet.nrows
118-
self.col_size = self.sheet.ncols
113+
if len(filename) > 0:
114+
self.excel = xlrd.open_workbook(filename)
115+
self.sheet_names = self.excel.sheet_names()
116+
self.sheet = self.excel.sheet_by_name(sheet_name) \
117+
if sheet_name is not None else self.excel.sheet_by_index(sheet_index)
118+
self.row_size = self.sheet.nrows
119+
self.col_size = self.sheet.ncols
119120

120-
def modify(self, sheet_index: int = None, sheet_name: str = None):
121+
def modify(self, sheet_index: int = None, sheet_name: str = None, file_name:str = None):
121122
"""
122123
修改初始化参数
124+
:param file_name: 文件名
123125
:param sheet_index: 工作簿序号
124126
:param sheet_name: 工作簿名称,这个优先级最高
125127
:return:
126128
"""
129+
self.filename = file_name if file_name is not None else self.filename
130+
if len(self.filename) > 0:
131+
self.excel = xlrd.open_workbook(self.filename)
132+
self.sheet_names = self.excel.sheet_names()
133+
self.sheet = self.excel.sheet_by_name(sheet_name) \
134+
if sheet_name is not None else self.excel.sheet_by_index(sheet_index)
135+
self.row_size = self.sheet.nrows
136+
self.col_size = self.sheet.ncols
127137
self.sheet_index = sheet_index if sheet_index is not None else self.sheet_index
128138
self.sheet_name = sheet_name if sheet_name is not None else self.sheet_name
129139
if sheet_index is not None or sheet_name is not None:

src/PaperCrawlerUtil/proxypool/crawlers/base.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22
import requests
33
from loguru import logger
44
from constant import *
5-
import global_val
5+
import PaperCrawlerUtil.global_val
66
from fake_headers import Headers
77
import time
88

src/PaperCrawlerUtil/proxypool/processors/getter.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
from loguru import logger
22

3-
import global_val
3+
import PaperCrawlerUtil.global_val
44
from proxypool.storages.redis import RedisClient
55
from proxypool.crawlers import __all__ as crawlers_cls
66
from proxypool.storages.proxy_dict import ProxyDict

src/PaperCrawlerUtil/proxypool/processors/server.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
from flask import Flask, g
22

3-
import global_val
3+
import PaperCrawlerUtil.global_val
44
from proxypool.exceptions import PoolEmptyException
55
from proxypool.storages.redis import RedisClient
66
from global_val import *

src/PaperCrawlerUtil/proxypool/processors/tester.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@
44
import aiohttp
55
from loguru import logger
66

7-
import global_val
7+
import PaperCrawlerUtil.global_val
88
from proxypool.schemas import Proxy
99
from proxypool.storages.redis import RedisClient
1010
from aiohttp import ClientProxyConnectionError, ServerDisconnectedError, ClientOSError, ClientHttpProxyError

src/PaperCrawlerUtil/proxypool/scheduler.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22
import multiprocessing
33
import sys
44

5-
import global_val
5+
import PaperCrawlerUtil.global_val
66

77
sys.path.append("../../PaperCrawlerUtil")
88
from proxypool.processors.server import app

src/PaperCrawlerUtil/proxypool/storages/proxy_dict.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@
1515
from loguru import logger
1616
from proxypool.utils.proxy import is_valid_proxy, convert_proxy_or_proxies
1717
from global_val import *
18-
import global_val
18+
import PaperCrawlerUtil.global_val
1919
from constant import *
2020

2121

src/PaperCrawlerUtil/proxypool/storages/redis.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
import redis
22

3-
import global_val
3+
import PaperCrawlerUtil.global_val
44
from proxypool.exceptions import PoolEmptyException
55
from proxypool.schemas import Proxy
66
from random import choice

0 commit comments

Comments
 (0)