活动介绍

D:\Users\lenovo\anaconda3\python.exe D:/pythonProject2/main.py Traceback (most recent call last): File "D:\pythonProject2\main.py", line 30, in <module> X = data.drop('day_28_flg', axis=1) # 特征列,删除表中为标签的那一列数据 AttributeError: 'numpy.ndarray' object has no attribute 'drop'

时间: 2023-11-27 07:03:09 浏览: 139
这个错误是由于你尝试在 numpy.ndarray 对象上调用 pandas 的 drop() 方法引起的。因为 numpy.ndarray 对象没有 drop() 方法,所以会发生 AttributeError 错误。 解决这个问题的方法是将 numpy.ndarray 对象转换为 pandas DataFrame 对象。你可以使用 pandas 的 DataFrame() 方法将 numpy 数组转换为 DataFrame。 例如,将 X 转换为 DataFrame 的代码如下: ``` import pandas as pd X = pd.DataFrame(X) ``` 这样你就可以在 DataFrame 上使用 drop() 方法了。
相关问题

D:\Users\lenovo\anaconda3\python.exe D:/pythonProject1/main.py Traceback (most recent call last): File "D:\pythonProject1\main.py", line 73, in <module> from imblearn.metrics import roc_curve_ovr ImportError: cannot import name 'roc_curve_ovr' from 'imblearn.metrics' (D:\Users\lenovo\anaconda3\lib\site-packages\imblearn\metrics\__init__.py)

这个错误提示表明,在导入imblearn.metrics模块的时候,无法找到名为'roc_curve_ovr'的函数或类。这可能是由于以下原因导致的: 1. 版本问题:在较老版本的imblearn.metrics中可能没有roc_curve_ovr函数或类,或者在最新版本中可能已经更改了名称或删除了该函数或类。 2. 安装问题:可能没有正确安装imblearn.metrics或者安装了不兼容的版本。 为了解决这个问题,你可以尝试以下步骤: 1. 检查imblearn.metrics中是否存在名为'roc_curve_ovr'的函数或类,并确认你的代码中正确地导入了该函数或类。 2. 升级imblearn.metrics到最新版本,并确认你的代码中正确地导入了该函数或类。 3. 确认你的Python环境中已经正确地安装了imblearn.metrics,并且安装了与你的Python版本和其他依赖项兼容的版本。 如果你仍然无法解决问题,请提供更多的上下文信息,例如你的代码和环境配置,我会尽力帮助你解决问题。

D:\pythonProject5\main.py:176: MatplotlibDeprecationWarning: Support for FigureCanvases without a required_interactive_framework attribute was deprecated in Matplotlib 3.6 and will be removed two minor releases later. plt.subplot(2, 5, i+1) D:\pythonProject5\main.py:178: MatplotlibDeprecationWarning: Support for FigureCanvases without a required_interactive_framework attribute was deprecated in Matplotlib 3.6 and will be removed two minor releases later. plt.show(ecg_signal) Traceback (most recent call last): File "D:\pythonProject5\main.py", line 219, in <module> main() File "D:\pythonProject5\main.py", line 178, in main plt.show(ecg_signal) File "D:\Users\lenovo\anaconda3\lib\site-packages\matplotlib\pyplot.py", line 446, in show return _get_backend_mod().show(*args, **kwargs) TypeError: Show.__call__() takes 1 positional argument but 2 were given

这个错误是由于`plt.show()`函数的调用方式不正确导致的。`plt.show()`函数不需要传入参数,它会自动显示当前所有的绘图结果。您可以将代码修改为以下形式: ``` python from matplotlib import pyplot as plt # 绘制ECG信号,并在一张图片上展示多个画布 fig = plt.figure(figsize=(20, 10)) for i in range(10): ax = fig.add_subplot(2, 5, i+1) ax.plot(ecg_signal[i]) ax.set_title(f'ECG Signal {i+1}') plt.show() ``` 这里的`plt.show()`函数不需要传入参数,它会显示`fig`中的所有画布。如果您需要在多个画布中显示多个信号,可以使用`fig.add_subplot()`函数来创建子画布,并在每个子画布上绘制对应的信号。
阅读全文

相关推荐

D:\Users\lenovo\anaconda3\envs\yolo8\python.exe D:/桌面/desktop/代码/ultralytics-main/1.py Traceback (most recent call last): File "D:\桌面\desktop\代码\ultralytics-main\ultralytics\nn\tasks.py", line 832, in torch_safe_load ckpt = torch.load(file, map_location="cpu") File "D:\桌面\desktop\代码\ultralytics-main\ultralytics\utils\patches.py", line 86, in torch_load return _torch_load(*args, **kwargs) File "D:\Users\lenovo\anaconda3\envs\yolo8\lib\site-packages\torch\serialization.py", line 1471, in load return _load( File "D:\Users\lenovo\anaconda3\envs\yolo8\lib\site-packages\torch\serialization.py", line 1964, in _load result = unpickler.load() File "D:\Users\lenovo\anaconda3\envs\yolo8\lib\site-packages\torch\serialization.py", line 1953, in find_class return super().find_class(mod_name, name) ModuleNotFoundError: No module named 'models' The above exception was the direct cause of the following exception: Traceback (most recent call last): File "D:\桌面\desktop\代码\ultralytics-main\1.py", line 4, in <module> model = YOLO(r"D:\桌面\desktop\代码\ultralytics-main\bestquanzhong\bestv7.pt") File "D:\桌面\desktop\代码\ultralytics-main\ultralytics\models\yolo\model.py", line 23, in __init__ super().__init__(model=model, task=task, verbose=verbose) File "D:\桌面\desktop\代码\ultralytics-main\ultralytics\engine\model.py", line 148, in __init__ self._load(model, task=task) File "D:\桌面\desktop\代码\ultralytics-main\ultralytics\engine\model.py", line 291, in _load self.model, self.ckpt = attempt_load_one_weight(weights) File "D:\桌面\desktop\代码\ultralytics-main\ultralytics\nn\tasks.py", line 905, in attempt_load_one_weight ckpt, weight = torch_safe_load(weight) # load ckpt File "D:\桌面\desktop\代码\ultralytics-main\ultralytics\nn\tasks.py", line 836, in torch_safe_load raise TypeError( TypeError: ERROR D:\\desktop\\ultralytics-main\bestquanzhong\bestv7.pt appears to be an Ultralytics YOLOv5 model originally trained with https://github.com

(scrapy_env) C:\Users\Lenovo\nepu_spider>scrapy crawl nepu 2025-07-04 11:44:20 [scrapy.utils.log] INFO: Scrapy 2.8.0 started (bot: nepu_spider) 2025-07-04 11:44:20 [scrapy.utils.log] INFO: Versions: lxml 4.9.3.0, libxml2 2.10.4, cssselect 1.1.0, parsel 1.6.0, w3lib 1.21.0, Twisted 22.10.0, Python 3.11.5 | packaged by Anaconda, Inc. | (main, Sep 11 2023, 13:26:23) [MSC v.1916 64 bit (AMD64)], pyOpenSSL 23.2.0 (OpenSSL 3.0.10 1 Aug 2023), cryptography 41.0.3, Platform Windows-10-10.0.26100-SP0 2025-07-04 11:44:20 [scrapy.crawler] INFO: Overridden settings: {'BOT_NAME': 'nepu_spider', 'FEED_EXPORT_ENCODING': 'utf-8', 'NEWSPIDER_MODULE': 'nepu_spider.spiders', 'REQUEST_FINGERPRINTER_IMPLEMENTATION': '2.7', 'SPIDER_MODULES': ['nepu_spider.spiders'], 'TWISTED_REACTOR': 'twisted.internet.asyncioreactor.AsyncioSelectorReactor'} 2025-07-04 11:44:20 [asyncio] DEBUG: Using selector: SelectSelector 2025-07-04 11:44:20 [scrapy.utils.log] DEBUG: Using reactor: twisted.internet.asyncioreactor.AsyncioSelectorReactor 2025-07-04 11:44:20 [scrapy.utils.log] DEBUG: Using asyncio event loop: asyncio.windows_events._WindowsSelectorEventLoop 2025-07-04 11:44:20 [scrapy.extensions.telnet] INFO: Telnet Password: 97bca17b548b5608 2025-07-04 11:44:20 [scrapy.middleware] INFO: Enabled extensions: ['scrapy.extensions.corestats.CoreStats', 'scrapy.extensions.telnet.TelnetConsole', 'scrapy.extensions.logstats.LogStats'] 2025-07-04 11:44:20 [scrapy.middleware] INFO: Enabled downloader middlewares: ['scrapy.downloadermiddlewares.httpauth.HttpAuthMiddleware', 'scrapy.downloadermiddlewares.downloadtimeout.DownloadTimeoutMiddleware', 'scrapy.downloadermiddlewares.defaultheaders.DefaultHeadersMiddleware', 'scrapy.downloadermiddlewares.useragent.UserAgentMiddleware', 'scrapy.downloadermiddlewares.retry.RetryMiddleware', 'scrapy.downloadermiddlewares.redirect.MetaRefreshMiddleware', 'scrapy.downloadermiddlewares.httpcompression.HttpCompressionMiddleware', 'scrapy.downloadermiddlewares.redirect.RedirectMiddleware', 'scrapy.downloadermiddlewares.cookies.CookiesMiddleware', 'scrapy.downloadermiddlewares.httpproxy.HttpProxyMiddleware', 'scrapy.downloadermiddlewares.stats.DownloaderStats'] 2025-07-04 11:44:20 [scrapy.middleware] INFO: Enabled spider middlewares: ['scrapy.spidermiddlewares.httperror.HttpErrorMiddleware', 'scrapy.spidermiddlewares.offsite.OffsiteMiddleware', 'scrapy.spidermiddlewares.referer.RefererMiddleware', 'scrapy.spidermiddlewares.urllength.UrlLengthMiddleware', 'scrapy.spidermiddlewares.depth.DepthMiddleware'] 2025-07-04 11:44:20 [scrapy.middleware] INFO: Enabled item pipelines: ['nepu_spider.pipelines.NepuSpiderPipeline'] 2025-07-04 11:44:20 [scrapy.core.engine] INFO: Spider opened 2025-07-04 11:44:21 [nepu] INFO: 🆕 数据表 NewsArticles 创建成功或已存在 2025-07-04 11:44:21 [nepu] INFO: ✅ 成功连接到 SQL Server 数据库 2025-07-04 11:44:21 [scrapy.extensions.logstats] INFO: Crawled 0 pages (at 0 pages/min), scraped 0 items (at 0 items/min) 2025-07-04 11:44:21 [scrapy.extensions.telnet] INFO: Telnet console listening on 127.0.0.1:6023 2025-07-04 11:44:21 [scrapy.core.engine] DEBUG: Crawled (200) <GET https://news.nepu.edu.cn/xsdt.htm> (referer: None) 2025-07-04 11:44:21 [scrapy.downloadermiddlewares.redirect] DEBUG: Redirecting (302) to <GET https://www.nepu.edu.cn/info/1049/9837.htm> from <GET http://www.nepu.edu.cn/info/1049/9837.htm> 2025-07-04 11:44:21 [scrapy.downloadermiddlewares.redirect] DEBUG: Redirecting (302) to <GET https://www.nepu.edu.cn/info/1049/9836.htm> from <GET http://www.nepu.edu.cn/info/1049/9836.htm> 2025-07-04 11:44:21 [scrapy.downloadermiddlewares.redirect] DEBUG: Redirecting (302) to <GET https://www.nepu.edu.cn/info/1049/9812.htm> from <GET http://www.nepu.edu.cn/info/1049/9812.htm> 2025-07-04 11:44:21 [scrapy.downloadermiddlewares.redirect] DEBUG: Redirecting (302) to <GET https://www.nepu.edu.cn/info/1049/9815.htm> from <GET http://www.nepu.edu.cn/info/1049/9815.htm> 2025-07-04 11:44:21 [scrapy.downloadermiddlewares.redirect] DEBUG: Redirecting (302) to <GET https://www.nepu.edu.cn/info/1049/9809.htm> from <GET http://www.nepu.edu.cn/info/1049/9809.htm> 2025-07-04 11:44:21 [scrapy.downloadermiddlewares.redirect] DEBUG: Redirecting (302) to <GET https://www.nepu.edu.cn/info/1049/9808.htm> from <GET http://www.nepu.edu.cn/info/1049/9808.htm> 2025-07-04 11:44:21 [scrapy.downloadermiddlewares.redirect] DEBUG: Redirecting (302) to <GET https://www.nepu.edu.cn/info/1049/10155.htm> from <GET http://www.nepu.edu.cn/info/1049/10155.htm> 2025-07-04 11:44:21 [scrapy.downloadermiddlewares.redirect] DEBUG: Redirecting (302) to <GET https://www.nepu.edu.cn/info/1049/10129.htm> from <GET http://www.nepu.edu.cn/info/1049/10129.htm> 2025-07-04 11:44:21 [scrapy.downloadermiddlewares.redirect] DEBUG: Redirecting (302) to <GET https://www.nepu.edu.cn/info/1049/9813.htm> from <GET http://www.nepu.edu.cn/info/1049/9813.htm> 2025-07-04 11:44:21 [scrapy.downloadermiddlewares.redirect] DEBUG: Redirecting (302) to <GET https://www.nepu.edu.cn/info/1049/10162.htm> from <GET http://www.nepu.edu.cn/info/1049/10162.htm> 2025-07-04 11:44:21 [scrapy.core.engine] DEBUG: Crawled (200) <GET https://www.nepu.edu.cn/info/1049/9812.htm> (referer: None) 2025-07-04 11:44:21 [scrapy.core.engine] DEBUG: Crawled (200) <GET https://www.nepu.edu.cn/info/1049/9809.htm> (referer: None) 2025-07-04 11:44:21 [scrapy.core.engine] DEBUG: Crawled (200) <GET https://www.nepu.edu.cn/info/1049/9837.htm> (referer: None) 2025-07-04 11:44:21 [scrapy.core.engine] DEBUG: Crawled (200) <GET https://www.nepu.edu.cn/info/1049/10155.htm> (referer: None) 2025-07-04 11:44:21 [scrapy.core.engine] DEBUG: Crawled (200) <GET https://www.nepu.edu.cn/info/1049/9815.htm> (referer: None) 2025-07-04 11:44:21 [scrapy.core.engine] DEBUG: Crawled (200) <GET https://www.nepu.edu.cn/info/1049/9813.htm> (referer: None) 2025-07-04 11:44:21 [scrapy.core.engine] DEBUG: Crawled (200) <GET https://www.nepu.edu.cn/info/1049/9836.htm> (referer: None) 2025-07-04 11:44:21 [scrapy.core.engine] DEBUG: Crawled (200) <GET https://www.nepu.edu.cn/info/1049/10162.htm> (referer: None) 2025-07-04 11:44:21 [scrapy.core.engine] DEBUG: Crawled (200) <GET https://www.nepu.edu.cn/info/1049/10129.htm> (referer: None) 2025-07-04 11:44:21 [scrapy.core.engine] DEBUG: Crawled (200) <GET https://www.nepu.edu.cn/info/1049/9808.htm> (referer: None) 2025-07-04 11:44:21 [scrapy.core.scraper] ERROR: Spider error processing <GET https://www.nepu.edu.cn/info/1049/9812.htm> (referer: None) Traceback (most recent call last): File "D:\annaCONDA\Lib\site-packages\scrapy\utils\defer.py", line 257, in iter_errback yield next(it) ^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\utils\python.py", line 312, in __next__ return next(self.data) ^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\utils\python.py", line 312, in __next__ return next(self.data) ^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\offsite.py", line 28, in <genexpr> return (r for r in result or () if self._filter(r, spider)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\referer.py", line 353, in <genexpr> return (self._set_referer(r, response) for r in result or ()) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\urllength.py", line 27, in <genexpr> return (r for r in result or () if self._filter(r, spider)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\depth.py", line 31, in <genexpr> return (r for r in result or () if self._filter(r, response, spider)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "C:\Users\Lenovo\nepu_spider\nepu_spider\spiders\nepu.py", line 33, in parse_detail text = response.css(selector).get() ^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\http\response\text.py", line 147, in css return self.selector.css(query) ^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\parsel\selector.py", line 282, in css return self.xpath(self._css2xpath(query)) ^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\parsel\selector.py", line 285, in _css2xpath return self._csstranslator.css_to_xpath(query) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\parsel\csstranslator.py", line 107, in css_to_xpath return super(HTMLTranslator, self).css_to_xpath(css, prefix) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\xpath.py", line 192, in css_to_xpath for selector in parse(css)) ^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 415, in parse return list(parse_selector_group(stream)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 428, in parse_selector_group yield Selector(*parse_selector(stream)) ^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 436, in parse_selector result, pseudo_element = parse_simple_selector(stream) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 544, in parse_simple_selector raise SelectorSyntaxError( cssselect.parser.SelectorSyntaxError: Expected selector, got <DELIM '/' at 0> 2025-07-04 11:44:21 [scrapy.core.scraper] ERROR: Spider error processing <GET https://www.nepu.edu.cn/info/1049/9809.htm> (referer: None) Traceback (most recent call last): File "D:\annaCONDA\Lib\site-packages\scrapy\utils\defer.py", line 257, in iter_errback yield next(it) ^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\utils\python.py", line 312, in __next__ return next(self.data) ^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\utils\python.py", line 312, in __next__ return next(self.data) ^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\offsite.py", line 28, in <genexpr> return (r for r in result or () if self._filter(r, spider)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\referer.py", line 353, in <genexpr> return (self._set_referer(r, response) for r in result or ()) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\urllength.py", line 27, in <genexpr> return (r for r in result or () if self._filter(r, spider)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\depth.py", line 31, in <genexpr> return (r for r in result or () if self._filter(r, response, spider)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "C:\Users\Lenovo\nepu_spider\nepu_spider\spiders\nepu.py", line 33, in parse_detail text = response.css(selector).get() ^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\http\response\text.py", line 147, in css return self.selector.css(query) ^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\parsel\selector.py", line 282, in css return self.xpath(self._css2xpath(query)) ^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\parsel\selector.py", line 285, in _css2xpath return self._csstranslator.css_to_xpath(query) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\parsel\csstranslator.py", line 107, in css_to_xpath return super(HTMLTranslator, self).css_to_xpath(css, prefix) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\xpath.py", line 192, in css_to_xpath for selector in parse(css)) ^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 415, in parse return list(parse_selector_group(stream)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 428, in parse_selector_group yield Selector(*parse_selector(stream)) ^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 436, in parse_selector result, pseudo_element = parse_simple_selector(stream) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 544, in parse_simple_selector raise SelectorSyntaxError( cssselect.parser.SelectorSyntaxError: Expected selector, got <DELIM '/' at 0> 2025-07-04 11:44:21 [scrapy.core.scraper] ERROR: Spider error processing <GET https://www.nepu.edu.cn/info/1049/9837.htm> (referer: None) Traceback (most recent call last): File "D:\annaCONDA\Lib\site-packages\scrapy\utils\defer.py", line 257, in iter_errback yield next(it) ^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\utils\python.py", line 312, in __next__ return next(self.data) ^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\utils\python.py", line 312, in __next__ return next(self.data) ^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\offsite.py", line 28, in <genexpr> return (r for r in result or () if self._filter(r, spider)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\referer.py", line 353, in <genexpr> return (self._set_referer(r, response) for r in result or ()) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\urllength.py", line 27, in <genexpr> return (r for r in result or () if self._filter(r, spider)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\depth.py", line 31, in <genexpr> return (r for r in result or () if self._filter(r, response, spider)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "C:\Users\Lenovo\nepu_spider\nepu_spider\spiders\nepu.py", line 33, in parse_detail text = response.css(selector).get() ^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\http\response\text.py", line 147, in css return self.selector.css(query) ^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\parsel\selector.py", line 282, in css return self.xpath(self._css2xpath(query)) ^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\parsel\selector.py", line 285, in _css2xpath return self._csstranslator.css_to_xpath(query) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\parsel\csstranslator.py", line 107, in css_to_xpath return super(HTMLTranslator, self).css_to_xpath(css, prefix) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\xpath.py", line 192, in css_to_xpath for selector in parse(css)) ^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 415, in parse return list(parse_selector_group(stream)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 428, in parse_selector_group yield Selector(*parse_selector(stream)) ^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 436, in parse_selector result, pseudo_element = parse_simple_selector(stream) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 544, in parse_simple_selector raise SelectorSyntaxError( cssselect.parser.SelectorSyntaxError: Expected selector, got <DELIM '/' at 0> 2025-07-04 11:44:21 [scrapy.core.scraper] ERROR: Spider error processing <GET https://www.nepu.edu.cn/info/1049/10155.htm> (referer: None) Traceback (most recent call last): File "D:\annaCONDA\Lib\site-packages\scrapy\utils\defer.py", line 257, in iter_errback yield next(it) ^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\utils\python.py", line 312, in __next__ return next(self.data) ^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\utils\python.py", line 312, in __next__ return next(self.data) ^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\offsite.py", line 28, in <genexpr> return (r for r in result or () if self._filter(r, spider)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\referer.py", line 353, in <genexpr> return (self._set_referer(r, response) for r in result or ()) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\urllength.py", line 27, in <genexpr> return (r for r in result or () if self._filter(r, spider)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\depth.py", line 31, in <genexpr> return (r for r in result or () if self._filter(r, response, spider)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "C:\Users\Lenovo\nepu_spider\nepu_spider\spiders\nepu.py", line 33, in parse_detail text = response.css(selector).get() ^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\http\response\text.py", line 147, in css return self.selector.css(query) ^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\parsel\selector.py", line 282, in css return self.xpath(self._css2xpath(query)) ^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\parsel\selector.py", line 285, in _css2xpath return self._csstranslator.css_to_xpath(query) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\parsel\csstranslator.py", line 107, in css_to_xpath return super(HTMLTranslator, self).css_to_xpath(css, prefix) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\xpath.py", line 192, in css_to_xpath for selector in parse(css)) ^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 415, in parse return list(parse_selector_group(stream)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 428, in parse_selector_group yield Selector(*parse_selector(stream)) ^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 436, in parse_selector result, pseudo_element = parse_simple_selector(stream) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 544, in parse_simple_selector raise SelectorSyntaxError( cssselect.parser.SelectorSyntaxError: Expected selector, got <DELIM '/' at 0> 2025-07-04 11:44:21 [scrapy.core.scraper] ERROR: Spider error processing <GET https://www.nepu.edu.cn/info/1049/9815.htm> (referer: None) Traceback (most recent call last): File "D:\annaCONDA\Lib\site-packages\scrapy\utils\defer.py", line 257, in iter_errback yield next(it) ^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\utils\python.py", line 312, in __next__ return next(self.data) ^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\utils\python.py", line 312, in __next__ return next(self.data) ^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\offsite.py", line 28, in <genexpr> return (r for r in result or () if self._filter(r, spider)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\referer.py", line 353, in <genexpr> return (self._set_referer(r, response) for r in result or ()) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\urllength.py", line 27, in <genexpr> return (r for r in result or () if self._filter(r, spider)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\depth.py", line 31, in <genexpr> return (r for r in result or () if self._filter(r, response, spider)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "C:\Users\Lenovo\nepu_spider\nepu_spider\spiders\nepu.py", line 33, in parse_detail text = response.css(selector).get() ^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\http\response\text.py", line 147, in css return self.selector.css(query) ^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\parsel\selector.py", line 282, in css return self.xpath(self._css2xpath(query)) ^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\parsel\selector.py", line 285, in _css2xpath return self._csstranslator.css_to_xpath(query) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\parsel\csstranslator.py", line 107, in css_to_xpath return super(HTMLTranslator, self).css_to_xpath(css, prefix) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\xpath.py", line 192, in css_to_xpath for selector in parse(css)) ^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 415, in parse return list(parse_selector_group(stream)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 428, in parse_selector_group yield Selector(*parse_selector(stream)) ^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 436, in parse_selector result, pseudo_element = parse_simple_selector(stream) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 544, in parse_simple_selector raise SelectorSyntaxError( cssselect.parser.SelectorSyntaxError: Expected selector, got <DELIM '/' at 0> 2025-07-04 11:44:21 [scrapy.core.scraper] ERROR: Spider error processing <GET https://www.nepu.edu.cn/info/1049/9813.htm> (referer: None) Traceback (most recent call last): File "D:\annaCONDA\Lib\site-packages\scrapy\utils\defer.py", line 257, in iter_errback yield next(it) ^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\utils\python.py", line 312, in __next__ return next(self.data) ^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\utils\python.py", line 312, in __next__ return next(self.data) ^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\offsite.py", line 28, in <genexpr> return (r for r in result or () if self._filter(r, spider)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\referer.py", line 353, in <genexpr> return (self._set_referer(r, response) for r in result or ()) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\urllength.py", line 27, in <genexpr> return (r for r in result or () if self._filter(r, spider)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\depth.py", line 31, in <genexpr> return (r for r in result or () if self._filter(r, response, spider)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "C:\Users\Lenovo\nepu_spider\nepu_spider\spiders\nepu.py", line 33, in parse_detail text = response.css(selector).get() ^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\http\response\text.py", line 147, in css return self.selector.css(query) ^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\parsel\selector.py", line 282, in css return self.xpath(self._css2xpath(query)) ^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\parsel\selector.py", line 285, in _css2xpath return self._csstranslator.css_to_xpath(query) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\parsel\csstranslator.py", line 107, in css_to_xpath return super(HTMLTranslator, self).css_to_xpath(css, prefix) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\xpath.py", line 192, in css_to_xpath for selector in parse(css)) ^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 415, in parse return list(parse_selector_group(stream)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 428, in parse_selector_group yield Selector(*parse_selector(stream)) ^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 436, in parse_selector result, pseudo_element = parse_simple_selector(stream) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 544, in parse_simple_selector raise SelectorSyntaxError( cssselect.parser.SelectorSyntaxError: Expected selector, got <DELIM '/' at 0> 2025-07-04 11:44:21 [scrapy.core.scraper] ERROR: Spider error processing <GET https://www.nepu.edu.cn/info/1049/9836.htm> (referer: None) Traceback (most recent call last): File "D:\annaCONDA\Lib\site-packages\scrapy\utils\defer.py", line 257, in iter_errback yield next(it) ^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\utils\python.py", line 312, in __next__ return next(self.data) ^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\utils\python.py", line 312, in __next__ return next(self.data) ^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\offsite.py", line 28, in <genexpr> return (r for r in result or () if self._filter(r, spider)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\referer.py", line 353, in <genexpr> return (self._set_referer(r, response) for r in result or ()) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\urllength.py", line 27, in <genexpr> return (r for r in result or () if self._filter(r, spider)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\depth.py", line 31, in <genexpr> return (r for r in result or () if self._filter(r, response, spider)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "C:\Users\Lenovo\nepu_spider\nepu_spider\spiders\nepu.py", line 33, in parse_detail text = response.css(selector).get() ^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\http\response\text.py", line 147, in css return self.selector.css(query) ^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\parsel\selector.py", line 282, in css return self.xpath(self._css2xpath(query)) ^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\parsel\selector.py", line 285, in _css2xpath return self._csstranslator.css_to_xpath(query) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\parsel\csstranslator.py", line 107, in css_to_xpath return super(HTMLTranslator, self).css_to_xpath(css, prefix) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\xpath.py", line 192, in css_to_xpath for selector in parse(css)) ^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 415, in parse return list(parse_selector_group(stream)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 428, in parse_selector_group yield Selector(*parse_selector(stream)) ^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 436, in parse_selector result, pseudo_element = parse_simple_selector(stream) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 544, in parse_simple_selector raise SelectorSyntaxError( cssselect.parser.SelectorSyntaxError: Expected selector, got <DELIM '/' at 0> 2025-07-04 11:44:21 [scrapy.core.scraper] ERROR: Spider error processing <GET https://www.nepu.edu.cn/info/1049/10129.htm> (referer: None) Traceback (most recent call last): File "D:\annaCONDA\Lib\site-packages\scrapy\utils\defer.py", line 257, in iter_errback yield next(it) ^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\utils\python.py", line 312, in __next__ return next(self.data) ^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\utils\python.py", line 312, in __next__ return next(self.data) ^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\offsite.py", line 28, in <genexpr> return (r for r in result or () if self._filter(r, spider)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\referer.py", line 353, in <genexpr> return (self._set_referer(r, response) for r in result or ()) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\urllength.py", line 27, in <genexpr> return (r for r in result or () if self._filter(r, spider)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\depth.py", line 31, in <genexpr> return (r for r in result or () if self._filter(r, response, spider)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "C:\Users\Lenovo\nepu_spider\nepu_spider\spiders\nepu.py", line 33, in parse_detail text = response.css(selector).get() ^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\http\response\text.py", line 147, in css return self.selector.css(query) ^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\parsel\selector.py", line 282, in css return self.xpath(self._css2xpath(query)) ^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\parsel\selector.py", line 285, in _css2xpath return self._csstranslator.css_to_xpath(query) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\parsel\csstranslator.py", line 107, in css_to_xpath return super(HTMLTranslator, self).css_to_xpath(css, prefix) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\xpath.py", line 192, in css_to_xpath for selector in parse(css)) ^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 415, in parse return list(parse_selector_group(stream)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 428, in parse_selector_group yield Selector(*parse_selector(stream)) ^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 436, in parse_selector result, pseudo_element = parse_simple_selector(stream) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 544, in parse_simple_selector raise SelectorSyntaxError( cssselect.parser.SelectorSyntaxError: Expected selector, got <DELIM '/' at 0> 2025-07-04 11:44:21 [scrapy.core.scraper] ERROR: Spider error processing <GET https://www.nepu.edu.cn/info/1049/10162.htm> (referer: None) Traceback (most recent call last): File "D:\annaCONDA\Lib\site-packages\scrapy\utils\defer.py", line 257, in iter_errback yield next(it) ^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\utils\python.py", line 312, in __next__ return next(self.data) ^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\utils\python.py", line 312, in __next__ return next(self.data) ^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\offsite.py", line 28, in <genexpr> return (r for r in result or () if self._filter(r, spider)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\referer.py", line 353, in <genexpr> return (self._set_referer(r, response) for r in result or ()) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\urllength.py", line 27, in <genexpr> return (r for r in result or () if self._filter(r, spider)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\depth.py", line 31, in <genexpr> return (r for r in result or () if self._filter(r, response, spider)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "C:\Users\Lenovo\nepu_spider\nepu_spider\spiders\nepu.py", line 33, in parse_detail text = response.css(selector).get() ^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\http\response\text.py", line 147, in css return self.selector.css(query) ^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\parsel\selector.py", line 282, in css return self.xpath(self._css2xpath(query)) ^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\parsel\selector.py", line 285, in _css2xpath return self._csstranslator.css_to_xpath(query) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\parsel\csstranslator.py", line 107, in css_to_xpath return super(HTMLTranslator, self).css_to_xpath(css, prefix) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\xpath.py", line 192, in css_to_xpath for selector in parse(css)) ^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 415, in parse return list(parse_selector_group(stream)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 428, in parse_selector_group yield Selector(*parse_selector(stream)) ^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 436, in parse_selector result, pseudo_element = parse_simple_selector(stream) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 544, in parse_simple_selector raise SelectorSyntaxError( cssselect.parser.SelectorSyntaxError: Expected selector, got <DELIM '/' at 0> 2025-07-04 11:44:21 [scrapy.core.scraper] ERROR: Spider error processing <GET https://www.nepu.edu.cn/info/1049/9808.htm> (referer: None) Traceback (most recent call last): File "D:\annaCONDA\Lib\site-packages\scrapy\utils\defer.py", line 257, in iter_errback yield next(it) ^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\utils\python.py", line 312, in __next__ return next(self.data) ^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\utils\python.py", line 312, in __next__ return next(self.data) ^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\offsite.py", line 28, in <genexpr> return (r for r in result or () if self._filter(r, spider)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\referer.py", line 353, in <genexpr> return (self._set_referer(r, response) for r in result or ()) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\urllength.py", line 27, in <genexpr> return (r for r in result or () if self._filter(r, spider)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "D:\annaCONDA\Lib\site-packages\scrapy\spidermiddlewares\depth.py", line 31, in <genexpr> return (r for r in result or () if self._filter(r, response, spider)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\core\spidermw.py", line 104, in process_sync for r in iterable: File "C:\Users\Lenovo\nepu_spider\nepu_spider\spiders\nepu.py", line 33, in parse_detail text = response.css(selector).get() ^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\scrapy\http\response\text.py", line 147, in css return self.selector.css(query) ^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\parsel\selector.py", line 282, in css return self.xpath(self._css2xpath(query)) ^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\parsel\selector.py", line 285, in _css2xpath return self._csstranslator.css_to_xpath(query) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\parsel\csstranslator.py", line 107, in css_to_xpath return super(HTMLTranslator, self).css_to_xpath(css, prefix) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\xpath.py", line 192, in css_to_xpath for selector in parse(css)) ^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 415, in parse return list(parse_selector_group(stream)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 428, in parse_selector_group yield Selector(*parse_selector(stream)) ^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 436, in parse_selector result, pseudo_element = parse_simple_selector(stream) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\annaCONDA\Lib\site-packages\cssselect\parser.py", line 544, in parse_simple_selector raise SelectorSyntaxError( cssselect.parser.SelectorSyntaxError: Expected selector, got <DELIM '/' at 0> 2025-07-04 11:44:21 [scrapy.core.engine] INFO: Closing spider (finished) 2025-07-04 11:44:21 [nepu] INFO: 🔌 已安全关闭数据库连接 2025-07-04 11:44:21 [scrapy.statscollectors] INFO: Dumping Scrapy stats: {'downloader/request_bytes': 5100, 'downloader/request_count': 21, 'downloader/request_method_count/GET': 21, 'downloader/response_bytes': 93797, 'downloader/response_count': 21, 'downloader/response_status_count/200': 11, 'downloader/response_status_count/302': 10, 'elapsed_time_seconds': 0.502389, 'finish_reason': 'finished', 'finish_time': datetime.datetime(2025, 7, 4, 3, 44, 21, 531133), 'httpcompression/response_bytes': 251471, 'httpcompression/response_count': 11, 'log_count/DEBUG': 24, 'log_count/ERROR': 10, 'log_count/INFO': 13, 'request_depth_max': 1, 'response_received_count': 11, 'scheduler/dequeued': 21, 'scheduler/dequeued/memory': 21, 'scheduler/enqueued': 21, 'scheduler/enqueued/memory': 21, 'spider_exceptions/SelectorSyntaxError': 10, 'start_time': datetime.datetime(2025, 7, 4, 3, 44, 21, 28744)} 2025-07-04 11:44:21 [scrapy.core.engine] INFO: Spider closed (finished) (scrapy_env) C:\Users\Lenovo\nepu_spider>

C:\Users\lenovo\anaconda3\envs\radi\python.exe F:\PythonProject2\transform.py Traceback (most recent call last): File "F:\PythonProject2\transform.py", line 29, in <module> output_file = convert_xls_to_xlsx(input_file) File "F:\PythonProject2\transform.py", line 18, in convert_xls_to_xlsx df = pd.read_excel(input_path, engine='xlrd') File "C:\Users\lenovo\anaconda3\envs\radi\lib\site-packages\pandas\util\_decorators.py", line 311, in wrapper return func(*args, **kwargs) File "C:\Users\lenovo\anaconda3\envs\radi\lib\site-packages\pandas\io\excel\_base.py", line 364, in read_excel io = ExcelFile(io, storage_options=storage_options, engine=engine) File "C:\Users\lenovo\anaconda3\envs\radi\lib\site-packages\pandas\io\excel\_base.py", line 1233, in __init__ self._reader = self._engines[engine](self._io, storage_options=storage_options) File "C:\Users\lenovo\anaconda3\envs\radi\lib\site-packages\pandas\io\excel\_xlrd.py", line 25, in __init__ super().__init__(filepath_or_buffer, storage_options=storage_options) File "C:\Users\lenovo\anaconda3\envs\radi\lib\site-packages\pandas\io\excel\_base.py", line 420, in __init__ self.book = self.load_workbook(self.handles.handle) File "C:\Users\lenovo\anaconda3\envs\radi\lib\site-packages\pandas\io\excel\_xlrd.py", line 38, in load_workbook return open_workbook(file_contents=data) File "C:\Users\lenovo\anaconda3\envs\radi\lib\site-packages\xlrd\__init__.py", line 148, in open_workbook bk = book.open_workbook_xls( File "C:\Users\lenovo\anaconda3\envs\radi\lib\site-packages\xlrd\book.py", line 92, in open_workbook_xls biff_version = bk.getbof(XL_WORKBOOK_GLOBALS) File "C:\Users\lenovo\anaconda3\envs\radi\lib\site-packages\xlrd\book.py", line 1278, in getbof bof_error('Expected BOF record; found %r' % self.mem[savpos:savpos+8]) File "C:\Users\lenovo\anaconda3\envs\radi\lib\site-packages\xlrd\book.py", line 1272, in bof_error raise XLRDError('Unsupported format, or corrupt file: ' + msg) xlrd.biffh.XLRDError: Unsupported format, or corrupt file: Expected BOF record; found b'diagnost'

大家在看

recommend-type

经典算法(C语言)

经典算法(C语言)包括1.汉若塔 2.费式数列 3. 巴斯卡三角形 4.三色棋 5.老鼠走迷官(一) 6.老鼠走迷官(二) 7.骑士走棋盘 8.八皇后 9.八枚银币 10.生命游戏 11.字串核对 12.双色、三色河内塔 13.背包问题(Knapsack Problem) 14.蒙地卡罗法求 PI 15.Eratosthenes筛选求质数 16.超长整数运算(大数运算) 17.长 PI 18.最大公因数、最小公倍数、因式分解 19.完美数 20.阿姆斯壮数 等等
recommend-type

公开公开公开公开-openprotocol_specification 2.7

LY-WCS-2012-01-06-01 V 1.0 公开公开公开公开 产品名称:产品名称:产品名称:产品名称: WCS 系统简介系统简介系统简介系统简介-公开版公开版公开版公开版 共共共共 13 页页页页 WCSWCSWCSWCS 系统简介系统简介系统简介系统简介 ((((客户交流用客户交流用客户交流用客户交流用)))) 文文文文 档档档档 作作作作 者:者:者:者: 王 超 日期:日期:日期:日期:2012/01/06 开发开发开发开发/测试经理:测试经理:测试经理:测试经理: 程 达 日期:日期:日期:日期:2012/01/06 项项项项 目目目目 经经经经 理:理:理:理: 程 达 日期:日期:日期:日期:2012/01/06 文文文文 档档档档 编编编编 号:号:号:号: ___________ ___ LY-WCS-2012-01-06-01______________ 上海朗因智能科技有限公司上海朗因智能科技有限公司上海朗因智能科技有限公司上海朗因智能科技有限公司 版权所有版权所有版权所有版权所有 不得复制不得复制不得复制不得复制
recommend-type

服务质量管理-NGBOSS能力架构

服务质量管理 二级能力名称 服务质量管理 二级能力编号 CMCM.5.4 概述 监测、分析和控制客户感知的服务表现 相关子能力描述 能够主动的将网络性能数据通告给前端客服人员; 能够根据按照客户价值来划分的客户群来制定特殊的SLA指标; 能够为最有价值的核心客户群进行网络优化; 对于常规的维护问题,QoS能够由网元设备自动完成,比如,对于网络故障的自恢复能力和优先客户的使用权; 能够把潜在的网络问题与客户进行主动的沟通; 能够分析所有的服务使用的质量指标; 能够根据关键的服务质量指标检测与实际的差距,提出改进建议; Service request 服务请求---请求管理。 客户的分析和报告:对关闭的请求、用户联系和相关的报告进行分析。 Marketing collateral的散发和marketing Collateral 的散发后的线索跟踪
recommend-type

extjs6.2加SenchaCmd-6.5.3.6-windows-64bit

SenchaCmd-6.5.3.6-windows-64bit ext6.2.0gpl SenchaCmd-6.5.3.6-windows-64bit ext6.2.0gpl
recommend-type

MOXA UPort1110drvUSB转串口驱动

MOXA UPort1110drvUSB转串口驱动,解决没有com口的烦恼

最新推荐

recommend-type

rclone-browser-1.7.0-2.el8.tar.gz

# 适用操作系统:Centos8 #Step1、解压 tar -zxvf xxx.el8.tar.gz #Step2、进入解压后的目录,执行安装 sudo rpm -ivh *.rpm
recommend-type

qt5pas-2.6-2001201.el8.tar.gz

# 适用操作系统:Centos8 #Step1、解压 tar -zxvf xxx.el8.tar.gz #Step2、进入解压后的目录,执行安装 sudo rpm -ivh *.rpm
recommend-type

TBS v3.7.0:支持PHP5的PHP模板引擎发布

根据给定文件信息,以下是关于“TinyButStrong v3.7.0 PHP模板引擎”的详细知识点介绍: 标题:“TinyButStrong v3.7.0 PHP模板引擎” 知识点: 1. 模板引擎定义与作用: 模板引擎是一种软件组件,它将页面的业务逻辑与显示逻辑分离,允许开发者在不修改服务器端代码的情况下,通过模板文件控制页面的布局和内容。模板引擎在Web开发中广泛应用,可以简化页面设计,提高代码的可维护性。 2. TinyButStrong (TBS)模板引擎: TBS是一个PHP模板引擎,它允许PHP脚本与HTML文件实现有效的分离。这种分离的好处是,网页设计师可以专注于模板设计,而不会触及后端的PHP代码,反之亦然。这使得团队合作变得更加顺畅,也降低了代码错误的可能性。 3. TBS的使用方式: TinyButStrong设计成可以兼容任何可视化HTML编辑器,如Adobe Dreamweaver、Microsoft Expression Web等,这意味着开发者可以使用这些工具进行模板设计,然后将这些模板嵌入到PHP脚本中。 4. TBS新版本特性: 该版本的TBS提供了6个新特性,虽然文件信息中没有具体列出这些特性,但一般来说,新特性可能涉及性能改进、新增函数、更灵活的模板操作、更加人性化的API设计等。 5. TBS版本修复的bug: 修复了5个bug,虽然没有具体说明,但一般修复bug会改善软件的稳定性和性能,提供更流畅的用户体验。 6. PHP版本支持变更: TBS v3.7.0只支持PHP5,不再支持PHP4。这一变更意味着TBS的使用环境将只限于运行PHP5及以上版本的服务器上。对于还在使用PHP4的开发者来说,需要升级到PHP5或以上版本才能继续使用TBS。 7. 内置对PDO和MySQLi的支持: PDO (PHP Data Objects)和MySQLi是PHP中用于数据库操作的两种扩展。PDO提供了一个数据访问抽象层,这意味着无论使用哪种数据库,都可以用相同的函数执行查询和获取数据。MySQLi是专门为MySQL数据库设计的面向对象的接口,它同样提供了丰富的功能来执行SQL语句和操作数据库。TBS内置对这两种数据库操作接口的支持,使得与数据库的交互变得更为简单和高效。 描述:“TinyButStrong(TBS)是一个可以让PHP脚本和HTML文件干净分离的PHP模板引擎。TBS设计成可以使用任何可视化HTML编辑器来编写TBS的HTML模板页。” 知识点: 1. TBS模板引擎的分离理念: 在Web开发中,将PHP脚本逻辑与HTML设计分离是非常重要的。这样做可以增强项目的可维护性,并允许网页设计师和后端开发人员独立工作,提高开发效率。TBS作为一个模板引擎,正是通过模板文件这一中间层实现了这种分离。 2. 可视化HTML编辑器的兼容性: TBS能够与各种可视化HTML编辑器相兼容,这一点极大地便利了前端开发人员。他们无需熟悉复杂的PHP代码,就可以设计出符合后端需要的模板。这样既保证了设计的自由度,也简化了前端和后端的协作流程。 标签:“PHP源码-其它类别” 知识点: 1. PHP源码的分类: PHP源码通常可以根据其功能和用途被分为不同的类别,如框架、库、模板引擎等。TBS作为模板引擎,归入了“其他类别”中,意味着它不属于框架或者常见的库类别,而是作为一个独立的工具来处理模板和数据的展示。 压缩包子文件的文件名称列表:“TinyButStrong_v3.7.0” 知识点: 1. 压缩包文件的命名规范: 通常情况下,压缩包文件的命名规范反映出其包含内容的版本信息、用途或者特性。在这里,“TinyButStrong_v3.7.0”明确指出了文件包含的软件名称和版本号,为用户提供了一个直观的参考。 2. 文件版本号的意义: 版本号通常表示了软件的迭代发展和更新程度。在本例中,“3.7.0”意味着这可能是第3个主版本的第7个次版本更新,版本号的递增通常伴随功能的增加、性能的提升或错误的修正。 综上所述,TinyButStrong v3.7.0作为一个PHP模板引擎,它在设计上注重了PHP脚本与HTML模板的分离,使得前端设计师可以独立于后端开发者工作,提高了工作效率。同时,它还提供了对最新PHP版本的支持,并内置了对PDO和MySQLi数据库操作的支持,使得数据操作更为便捷。新特性的加入和bug的修复也表明了TBS在不断的发展和优化中,以满足日益增长的Web开发需求。
recommend-type

【FFT算法的硬件实现探索】:FPGA与ASIC优化技术分析

# 摘要 本论文综合讨论了快速傅里叶变换(FFT)算法的基础知识、FPGA技术与ASIC技术在FFT实现中的应用,及其各自的优化技术。文中还对FFT硬
recommend-type

stm32stlinkutility擦除程序无法擦除写保护

<think>好的,我现在需要帮助用户解决STM32使用ST-LINK Utility时因写保护导致程序无法擦除的问题。首先,我应该回顾用户提供的引用资料,看看有没有相关的解决方案。 用户提到了引用[2]和[3],其中引用[2]提到了使用仿真器解除写保护的方法,而引用[3]则直接提到使用STM32 ST-LINK Utility去写保护的步骤。此外,引用[4]介绍了不同的烧录方式,可能也有帮助。 首先,用户的问题是无法擦除程序,因为芯片被写保护了。根据经验,STM32的写保护通常可以通过连接ST-LINK Utility后进入设置,解除保护。可能需要检查连接是否正常,然后通过软件操作来解除
recommend-type

C++程序设计郑莉教学视频与课件资源打包下载

根据提供的文件信息,我们可以推断出该压缩包内含的教学资源主要围绕C++语言程序设计,由郑莉老师制作,内容涉及教学视频、课件以及书籍的源代码。下面将详细展开这些知识点。 ### C++程序设计基础 C++是一种静态数据类型、编译式、通用的编程语言,它支持多种编程范式,包括过程化、面向对象和泛型编程。C++被设计为对硬件操作有强大的控制能力,同时拥有高度的代码复用性。作为C语言的超集,C++在继承C语言高效、紧凑等特性的同时,增加了面向对象编程的能力。 ### 郑莉与C++教学 郑莉老师制作的C++教学资源,可能包含了以下内容: 1. **教学视频(45集)**:这套视频很可能是分阶段的教学,由浅入深地介绍C++的基础知识到高级特性。集数的分布可能按照如下逻辑: - 基础语法、数据类型和变量 - 控制结构、循环和函数 - 数组、字符串操作和指针 - 结构体、联合体和枚举 - 面向对象编程基础:类和对象 - 继承、多态性和虚函数 - 静态成员、友元函数和运算符重载 - 高级特性:模板、异常处理、STL(标准模板库) - C++11/14/17/20新特性介绍 - 项目实战与案例分析 2. **课件(PPT)**:这个部分可能包含了对应上述视频的教学PPT,每集视频配套的PPT都会详细解释关键点,通常包括了课程大纲、重点、难点的图示和代码示例。PPT可能还会提供一些练习题和答案,帮助学习者加强理解和实践。 3. **源代码**:这里的源代码应该包括了教学视频中演示的示例代码以及课件中提及的代码片段。这些代码可能覆盖了从简单的“Hello World”程序到复杂的面向对象设计案例,以及各种C++的STL使用实例。源代码的提供有助于学习者能够动手实践,更好地理解和掌握C++编程。 ### C++学习资源与方法 1. **理论与实践相结合**:学习C++的过程中,理论知识的理解和实际编程操作需要相辅相成,特别是在面向对象编程方面,实际编写代码并观察运行结果,能极大加深理解。 2. **重视基础**:C++语言基础特别重要,特别是数据类型、运算符、控制流程、函数和指针等方面。这些都是学习后续内容(如类和对象)的基石。 3. **面向对象深入理解**:C++的核心是面向对象编程(OOP)。学习者需要透彻理解类与对象、封装、继承和多态的概念,并且通过实际编码体验OOP的优势。 4. **掌握STL**:标准模板库(STL)是C++语言的一个重要组成部分,它提供了一系列可重用的模板类和函数,涵盖容器、迭代器、算法、函数对象等,学会使用STL可以显著提高开发效率。 5. **关注新标准**:C++标准从C++98/03到C++11、C++14、C++17乃至C++20不断发展更新,每一次更新都引入了新的特性,如自动类型推导、lambda表达式、智能指针、并发支持等。学习新标准对从事现代C++开发至关重要。 6. **实战与项目经验**:学习编程最终是为了能够解决实际问题,通过项目实战可以锻炼综合运用所学知识的能力,并且了解软件开发的完整流程。 7. **社区资源**:积极参与C++社区,如Stack Overflow、Reddit等论坛,可以获取额外的学习资源、解决问题的思路和最新的行业动态。 ### 结语 通过上述对【标题】、【描述】和【压缩包子文件的文件名称列表】的分析,我们可以得知这个压缩包提供了全面的C++编程学习材料。无论对于初学者,还是有一定基础希望进一步提升的开发者,这份由郑莉老师制作的教学资源都具有很高的参考价值。通过系统地学习这些材料,并结合实践和项目经验,学习者可以有效提高C++编程能力。
recommend-type

【突破FFT算法速度瓶颈】:并行化实现与计算性能优化

# 摘要 本文全面综述了快速傅里叶变换(FFT)算法及其并行化实现,阐述了并行计算的理论基础,包括定义、重要性、模型以及算法设计基础。文章深入探讨FFT算法的并行化策略,并提供了基于不同并行计算平台(如MPI和OpenMP)的编程实践案例。此外,本文还对并行FFT算法的性能进行了分析,并评估了其在现代科学和工业应用中的潜力。通过实验环境和工具的介绍,文章对大规模数
recommend-type

SSL Error: Unable to verify the first certificate

<think>首先,用户的问题是:“SSL Error: Unable to verify the first certificate”。这是一个关于SSL证书错误的常见问题。我需要用中文回答这个问题。 用户指定:请在回答完问题后,必须根据用户的问题及你的回答提出3-5个 §§相关问题§§。第一行必须完整且输出 "§§相关问题§§"!禁止换行分散。 所以,我的回答结构应该是: 1. 回答用户的问题(用中文)。 2. 然后,以 "§§相关问题§§" 开头的一行。 3. 紧接着,列出3-5个相关问题,每个问题以数字点开头,如 "1. ···"。 现在,针对用户的问题:"SSL Error:
recommend-type

创意jQuery火箭动画404错误页教程

根据提供的文件信息,我们可以详细解释以下几个知识点: 1. jQuery技术基础 jQuery是一个快速、小巧、功能丰富的JavaScript库,其设计目的是使网页上常见的JavaScript任务更简单、更快速。它封装了JavaScript中常用的DOM操作、事件处理、动画效果和Ajax交互,使其以更简洁的方式编写。jQuery的核心是一个名为jQuery()的函数,这个函数可以接收一个选择器表达式,返回一个包含所有匹配元素的jQuery对象。这个对象可以被用来进行各种操作,包括改变样式、添加事件监听器、创建动画效果等。 2. 动态网页设计与动画实现 动态网页设计是指利用客户端或服务器端的技术创建具有交互性和动态内容的网页。动画效果是动态网页设计中非常重要的一环,它可以让网页更具吸引力,提供更加丰富的用户体验。jQuery是实现前端动画的常用工具之一,它提供了丰富的动画方法,例如:`fadeIn()`, `fadeOut()`, `slideToggle()`, `animate()`等。这些方法可以让我们以非常简单的方式实现各种动画效果,而无需深入了解复杂的JavaScript动画编程。 3. 404错误页面的定制化与用户体验 404错误页面是网站中出现的“找不到网页”的错误提示页面。虽然404页面本身表明用户访问了一个不存在的资源,但这并不意味着它就应该平淡无奇或者缺乏设计。事实上,一个好的404页面可以成为改善用户体验的工具。定制化的404页面可以提供有用的信息,比如提供返回首页的链接、搜索框、网站导航菜单,或者一些有趣的元素和信息,以减轻用户由于错误访问带来的挫败感。 4. 基于jQuery的火箭动画实现 从描述中我们知道,此404错误提示页使用jQuery技术实现了一个火箭动画效果。通过使用jQuery提供的动画方法,开发者可以创建一个火箭起飞并穿越屏幕的动画。具体实现时,开发者需要编写JavaScript代码,通过操作DOM元素的样式、位置和可见性来模拟动画效果。代码可能会涉及到定时器(如`setInterval`或`setTimeout`)来按时间间隔更新DOM元素,从而生成连续的画面变化,模拟动画效果。 5. 脚本资源的获取与利用 脚本资源是动态网页开发中不可或缺的一部分。在本例中,脚本资源特指使用jQuery库来实现特定功能的代码。由于jQuery是一个开源库,它的脚本资源可以很容易地从网上下载,并且通过CDN(内容分发网络)来加载到项目中。一旦加载了jQuery脚本,开发者就可以利用其封装好的方法来编写网页动画和交互。此外,通过阅读和分析其他开发者共享的脚本资源,比如本例中的火箭404错误页面,可以学习到动画实现的技巧和方法,这对于提升个人的前端开发技能是很有帮助的。 综上所述,本文件涉及到的知识点包括了jQuery技术的应用、动态网页设计和动画效果的实现、404错误页面的定制化设计以及脚本资源的使用和学习。通过学习这些知识点,可以提高前端开发的实践能力,并优化网页用户体验。
recommend-type

【高效FFT算法新实现】:一种快速处理数字信号的方法

# 摘要 本文对数字信号处理中的快速傅里叶变换(FFT)算法进行了全面的概述和分析。首先介绍了FFT算法的理论基础,包括离散傅里叶变换(DFT)的概念、FFT的历史重要性及其数学优化原理。接着,探讨了FFT算法的编程实现,涵盖了编程语言选择、算法框架设计、以及递归与迭代两种实现方法。本文还详细论述了FFT算法在频谱分