Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Contribute to GitLab
Sign in / Register
Toggle navigation
E
esi-table-data
Project
Project
Details
Activity
Cycle Analytics
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Issues
0
Issues
0
List
Board
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Charts
Wiki
Wiki
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
esi-data-scrapping
esi-table-data
Commits
6ddd556a
Commit
6ddd556a
authored
Jul 25, 2017
by
Vasyl Bodnaruk
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
Fix searching by tag
parent
87b3f20f
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
19 additions
and
21 deletions
+19
-21
aitop.py
exa/exa/spiders/aitop.py
+19
-21
No files found.
exa/exa/spiders/aitop.py
View file @
6ddd556a
...
...
@@ -48,40 +48,38 @@ class AitopSpider(BaseSpider):
else
:
return
False
for
i
in
self
.
build_items
(
response
):
items
,
is_duplicate
=
self
.
build_items
(
response
)
for
i
in
items
:
if
is_company_in_item
(
i
):
yield
i
next_url
=
self
.
next_url
(
response
)
if
next_url
:
if
self
.
can_follow
(
next_url
,
is_duplicate
)
:
yield
scrapy
.
Request
(
next_url
,
callback
=
self
.
parse_by_title_description
,
meta
=
response
.
meta
)
else
:
print
(
"DUPLICATE NEWS"
)
def
parse_by_tag
(
self
,
response
):
try
:
for
i
in
self
.
build_items
(
response
):
yield
i
next_url
=
self
.
next_url
(
response
)
if
next_url
:
yield
scrapy
.
Request
(
next_url
,
callback
=
self
.
parse_by_tag
,
meta
=
response
.
meta
)
except
:
pass
def
build_items
(
self
,
response
):
try
:
items
=
list
()
rows
=
response
.
xpath
(
".//div[contains(@class, 'summaries')]"
)
for
r
in
rows
:
i
=
r
.
xpath
(
"//div[@class='row']"
)
is_duplicate
=
False
for
i
in
rows
:
# i = r.xpath("//div[@class='row']")
item
=
ExaItem
()
item
[
'date'
]
=
dateparser
.
parse
(
i
.
xpath
(
".//time/@datetime"
)
.
extract_first
())
.
replace
(
tzinfo
=
None
)
item
[
'title'
]
=
''
.
join
(
i
.
xpath
(
".//div[contains(@class, 'col-xs-12')]/h3/a//text()"
)
.
extract
()
)
item
[
'description'
]
=
''
.
join
(
i
.
xpath
(
".//div[@class='summary-content']/p/text()"
)
.
extract
()
)
item
[
'date'
]
=
dateparser
.
parse
(
i
.
xpath
(
"./
div[@class='row']/
/time/@datetime"
)
.
extract_first
())
.
replace
(
tzinfo
=
None
)
item
[
'title'
]
=
i
.
xpath
(
"./div[@class='row']//div[contains(@class, 'col-xs-12')]//h3/a//text()"
)
.
extract
(
)
item
[
'description'
]
=
i
.
xpath
(
"./div[@class='row']//div[@class='summary-content']/p//text()"
)
.
extract
(
)
item
[
'url'
]
=
i
.
xpath
(
".//div[contains(@class, 'col-xs-12')]/h3/a/@href"
)
.
extract_first
()
item
.
update
(
self
.
get_common_items
(
response
.
meta
[
'company'
]))
item
[
'post_id'
]
=
response
.
meta
[
'post_id'
]
item
[
'tags'
]
=
r
.
xpath
(
".//div[@class='row hidden-xs']//div[@title='Concept Tags']//a[@class='filter btn btn-link']/text()"
)
.
extract
()
item
[
'tags'
]
=
i
.
xpath
(
".//div[@class='row hidden-xs']//div[@title='Concept Tags']//a[@class='filter btn btn-link']/text()"
)
.
extract
()
yield
item
next_url
=
self
.
next_url
(
response
)
if
next_url
:
print
(
'FOLLOW'
)
# yield scrapy.Request(next_url, callback=self.parse_by_tag, meta=response.meta)
items
.
append
(
item
)
return
items
except
:
pass
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment