| | |
| | | log.pages = spider.crawler.stats.get_value('response_received_count', 0) |
| | | log.items = spider.crawler.stats.get_value('item_scraped_count', 0) |
| | | log.finish_time=datetime.datetime.fromtimestamp(int(time.time())).strftime('%Y-%m-%d %H:%M:%S') |
| | | log.status=3 |
| | | self.session.commit() |
| | | except SQLAlchemyError as e: |
| | | self.session.rollback() |
| | |
| | | log.pages = spider.crawler.stats.get_value('response_received_count', 0) |
| | | log.items = spider.crawler.stats.get_value('item_scraped_count', 0) |
| | | log.finish_time=datetime.datetime.fromtimestamp(int(time.time())).strftime('%Y-%m-%d %H:%M:%S') |
| | | log.status=3 |
| | | self.session.commit() |
| | | except SQLAlchemyError as e: |
| | | self.session.rollback() |
| | |
| | | log.pages = spider.crawler.stats.get_value('response_received_count', 0) |
| | | log.items = spider.crawler.stats.get_value('item_scraped_count', 0) |
| | | log.finish_time=datetime.datetime.fromtimestamp(int(time.time())).strftime('%Y-%m-%d %H:%M:%S') |
| | | log.status=3 |
| | | self.session.commit() |
| | | except SQLAlchemyError as e: |
| | | self.session.rollback() |
| | |
| | | log.pages = spider.crawler.stats.get_value('response_received_count', 0) |
| | | log.items = spider.crawler.stats.get_value('item_scraped_count', 0) |
| | | log.finish_time=datetime.datetime.fromtimestamp(int(time.time())).strftime('%Y-%m-%d %H:%M:%S') |
| | | log.status=3 |
| | | self.session.commit() |
| | | except SQLAlchemyError as e: |
| | | self.session.rollback() |
| | |
| | | log.pages = spider.crawler.stats.get_value('response_received_count', 0) |
| | | log.items = spider.crawler.stats.get_value('item_scraped_count', 0) |
| | | log.finish_time=datetime.datetime.fromtimestamp(int(time.time())).strftime('%Y-%m-%d %H:%M:%S') |
| | | log.status=3 |
| | | self.session.commit() |
| | | except SQLAlchemyError as e: |
| | | self.session.rollback() |
| | |
| | | log.pages = spider.crawler.stats.get_value('response_received_count', 0) |
| | | log.items = spider.crawler.stats.get_value('item_scraped_count', 0) |
| | | log.finish_time=datetime.datetime.fromtimestamp(int(time.time())).strftime('%Y-%m-%d %H:%M:%S') |
| | | log.status=3 |
| | | self.session.commit() |
| | | except SQLAlchemyError as e: |
| | | self.session.rollback() |
| | |
| | | log.pages = spider.crawler.stats.get_value('response_received_count', 0) |
| | | log.items = spider.crawler.stats.get_value('item_scraped_count', 0) |
| | | log.finish_time=datetime.datetime.fromtimestamp(int(time.time())).strftime('%Y-%m-%d %H:%M:%S') |
| | | log.status=3 |
| | | self.session.commit() |
| | | except SQLAlchemyError as e: |
| | | self.session.rollback() |
| | |
| | | log.pages = spider.crawler.stats.get_value('response_received_count', 0) |
| | | log.items = spider.crawler.stats.get_value('item_scraped_count', 0) |
| | | log.finish_time=datetime.datetime.fromtimestamp(int(time.time())).strftime('%Y-%m-%d %H:%M:%S') |
| | | log.status=3 |
| | | self.session.commit() |
| | | except SQLAlchemyError as e: |
| | | self.session.rollback() |
| | |
| | | log.pages = spider.crawler.stats.get_value('response_received_count', 0) |
| | | log.items = spider.crawler.stats.get_value('item_scraped_count', 0) |
| | | log.finish_time=datetime.datetime.fromtimestamp(int(time.time())).strftime('%Y-%m-%d %H:%M:%S') |
| | | log.status=3 |
| | | self.session.commit() |
| | | except SQLAlchemyError as e: |
| | | self.session.rollback() |
| | |
| | | log.pages = spider.crawler.stats.get_value('response_received_count', 0) |
| | | log.items = spider.crawler.stats.get_value('item_scraped_count', 0) |
| | | log.finish_time=datetime.datetime.fromtimestamp(int(time.time())).strftime('%Y-%m-%d %H:%M:%S') |
| | | log.status=3 |
| | | self.session.commit() |
| | | except SQLAlchemyError as e: |
| | | self.session.rollback() |
| | |
| | | log.pages = spider.crawler.stats.get_value('response_received_count', 0) |
| | | log.items = spider.crawler.stats.get_value('item_scraped_count', 0) |
| | | log.finish_time=datetime.datetime.fromtimestamp(int(time.time())).strftime('%Y-%m-%d %H:%M:%S') |
| | | log.status=3 |
| | | self.session.commit() |
| | | except SQLAlchemyError as e: |
| | | self.session.rollback() |