@@ -91,6 +91,9 @@ def __init__(self, *a, **kw):
91
91
92
92
def start (self ):
93
93
self .start_crawling ()
94
+ self .start_reactor ()
95
+
96
+ def start_reactor (self ):
94
97
if self .settings .getbool ('DNSCACHE_ENABLED' ):
95
98
reactor .installResolver (CachingThreadedResolver (reactor ))
96
99
reactor .addSystemEventTrigger ('before' , 'shutdown' , self .stop )
@@ -102,7 +105,7 @@ def start_crawling(self):
102
105
def stop (self ):
103
106
raise NotImplementedError
104
107
105
- def _stop_reactor (self , _ = None ):
108
+ def stop_reactor (self , _ = None ):
106
109
try :
107
110
reactor .stop ()
108
111
except RuntimeError : # raised if already stopped or in shutdown stage
@@ -120,7 +123,7 @@ def _signal_kill(self, signum, _):
120
123
signame = signal_names [signum ]
121
124
log .msg (format = 'Received %(signame)s twice, forcing unclean shutdown' ,
122
125
level = log .INFO , signame = signame )
123
- reactor .callFromThread (self ._stop_reactor )
126
+ reactor .callFromThread (self .stop_reactor )
124
127
125
128
126
129
class CrawlerProcess (ProcessMixin ):
@@ -140,7 +143,7 @@ def create_crawler(self, name=None):
140
143
141
144
return self .crawlers [name ]
142
145
143
- def start_crawling (self ):
146
+ def start_crawler (self ):
144
147
name , crawler = self .crawlers .popitem ()
145
148
146
149
sflo = log .start_from_crawler (crawler )
@@ -157,17 +160,13 @@ def start_crawling(self):
157
160
158
161
def check_done (self , ** kwargs ):
159
162
if self .crawlers and not self .stopping :
160
- self .start_crawling ()
163
+ self .start_crawler ()
161
164
else :
162
- self ._stop_reactor ()
165
+ self .stop_reactor ()
163
166
164
- def print_headers (self ):
167
+ def start_crawling (self ):
165
168
log .scrapy_info (self .settings )
166
-
167
- def start (self , headers = True ):
168
- if headers :
169
- self .print_headers ()
170
- return super (CrawlerProcess , self ).start ()
169
+ self .start_crawler ()
171
170
172
171
@defer .inlineCallbacks
173
172
def stop (self ):
0 commit comments