为您找到相关结果303,190个
python多进程提取处理大量文本的关键词方法_python_脚本之家
def extract_keyword(input_string): #print("Do task by process {proc}".format(proc=os.getpid())) tags = jieba.analyse.extract_tags(input_string, topK=100) #print("key words:{kw}".format(kw=" ".join(tags))) return tags #def parallel_extract_keyword(input_string,out_file): def para...
www.jb51.net/article/1415...htm 2025-3-12
深入理解Python中的内置常量_python_脚本之家
1 2 >>> True = 1 SyntaxError: can't assign to keyword二. False 1. False是bool类型用来表示假值的常量。1 2 3 4 >>> False False >>> type(False) <class 'bool'>2. 对常量False进行任何赋值操作都会抛出语法错误。1 2 >>> False = 0 SyntaxError: can't assign to keyword...
www.jb51.net/article/1141...htm 2025-3-13
Python大批量搜索引擎图像爬虫工具详解_python_脚本之家
bing_crawler.crawl(keyword=word, max_num=2000) #百度爬虫 # baidu_storage = {'root_dir': 'baidu\\' + word} # baidu_crawler = BaiduImageCrawler(parser_threads=2, # downloader_threads=4, # storage=baidu_storage) # baidu_crawler.crawl(keyword=word, # max_num=2000) # google爬虫 # goog...
www.jb51.net/article/1997...htm 2025-2-24
Python 结巴分词实现关键词抽取分析_python_脚本之家
实现TF-IDF算法抽取关键词的类TFIDF和实现TextRank算法抽取关键词的类TextRank都是类KeywordExtractor的子类。而在类KeywordExtractor,实现了一个方法,可以根据用户指定的路径,加载用户提供的停用词集合。 类KeywordExtractor是在jieba/analyse/tfidf.py中实现。 类KeywordExtractor首先提供了一个默认的名为STOP_WORDS的...
www.jb51.net/article/1264...htm 2025-3-13
python爬虫之爬取百度翻译_python_脚本之家
keyword=input("需要翻译的词:") data={ "kw": keyword } # UA 伪装 header={ "user-agent":"Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/89.0.4389.90 Safari/537.36" } # 发送请求 response=requests.post(url=url, data=data, headers=header).json...
www.jb51.net/article/2107...htm 2025-3-12
python用ConfigObj读写配置文件的实现代码_python_脚本之家
config['keyword1'] = value1 config['keyword2'] = value2 # config['section1'] = {} config['section1']['keyword3'] = value3 config['section1']['keyword4'] = value4 # section2 = { 'keyword5': value5, 'keyword6': value6, ...
www.jb51.net/article/344...htm 2025-3-9
Django中使用Whoosh进行全文检索的方法_python_脚本之家
query = mp.parse(keyword) with indexer.searcher() as searcher: results = searcher.search(query, limit=15) articles = [] for hit in results: log.debug(hit) articles.append({ 'id': hit['id'], 'slug': hit['slug'], }) return articles def rebuild(): if os.path.exists(index_dir)...
www.jb51.net/article/1587...htm 2025-2-22