Commit e7af713d authored by delanoe's avatar delanoe

[FEAT] adding scrapers, maybe to add inside the API.

parent 15296ea7
...@@ -26,7 +26,7 @@ from annotations.views import main as annotations_main_view ...@@ -26,7 +26,7 @@ from annotations.views import main as annotations_main_view
from graphExplorer.rest import Graph from graphExplorer.rest import Graph
from graphExplorer.views import explorer from graphExplorer.views import explorer
from scrappers import urls as scrappers_urls from scrapers import urls as scrapers_urls
urlpatterns = [ url(r'^admin/', admin.site.urls) urlpatterns = [ url(r'^admin/', admin.site.urls)
, url(r'^generated/', include(gargantext.views.generated.urls)) , url(r'^generated/', include(gargantext.views.generated.urls))
...@@ -45,5 +45,5 @@ urlpatterns = [ url(r'^admin/', admin.site.urls) ...@@ -45,5 +45,5 @@ urlpatterns = [ url(r'^admin/', admin.site.urls)
, url(r'^projects/(\d+)/corpora/(\d+)/node_link.json$', Graph.as_view()) , url(r'^projects/(\d+)/corpora/(\d+)/node_link.json$', Graph.as_view())
#url(r'^projects/(\d+)/corpora/(\d+)/explorer$', include(graphExplorer.urls)) #url(r'^projects/(\d+)/corpora/(\d+)/explorer$', include(graphExplorer.urls))
#url(r'^projects/(\d+)/corpora/(\d+)/explorer$', include(graphExplorer_urls)) #url(r'^projects/(\d+)/corpora/(\d+)/explorer$', include(graphExplorer_urls))
, url(r'^scrappers/', include(scrappers_urls)) , url(r'^scrapers/', include(scrapers_urls))
] ]
...@@ -2,7 +2,6 @@ from django.conf.urls import url ...@@ -2,7 +2,6 @@ from django.conf.urls import url
from . import nodes from . import nodes
from . import ngramlists from . import ngramlists
from . import scraping
urlpatterns = [ url(r'^nodes$' , nodes.NodeListResource.as_view()) urlpatterns = [ url(r'^nodes$' , nodes.NodeListResource.as_view())
, url(r'^nodes/(\d+)$' , nodes.NodeResource.as_view() ) , url(r'^nodes/(\d+)$' , nodes.NodeResource.as_view() )
...@@ -22,7 +21,4 @@ urlpatterns = [ url(r'^nodes$' , nodes.NodeListResource.as_view() ...@@ -22,7 +21,4 @@ urlpatterns = [ url(r'^nodes$' , nodes.NodeListResource.as_view()
# aka lexical model # aka lexical model
#, url(r'^scraping$' , scraping.Target.as_view() )
# Scraping : getting data from external database
# Available databases : Pubmed, IsTex, (next: CERN)
] ]
from scrappers.MedlineFetcher import MedlineFetcher from scrapers.MedlineFetcher import MedlineFetcher
# from datetime import datetime # from datetime import datetime
...@@ -24,7 +24,7 @@ from gargantext.util.toolchain import parse_extract_indexhyperdata ...@@ -24,7 +24,7 @@ from gargantext.util.toolchain import parse_extract_indexhyperdata
# pour lire la section [scrappers] de gargantext.ini # pour lire la section [scrapers] de gargantext.ini
#from configparser import ConfigParser #from configparser import ConfigParser
# -------------------------------------------------------------------- # --------------------------------------------------------------------
......
from django.conf.urls import url from django.conf.urls import url
import scrappers.pubmed as pubmed import scrapers.pubmed as pubmed
#import scrapers.istex as istex
#import scrapers.cern as cern
#import scrapers.hal as hal
# /!\ urls patterns here are *without* the trailing slash # Scraping : getting data from external database
# Available databases : Pubmed, IsTex, (next: CERN)
# /!\ urls patterns here are *without* the trailing slash
urlpatterns = [ url(r'^pubmed/query$', pubmed.getGlobalStats) urlpatterns = [ url(r'^pubmed/query$', pubmed.getGlobalStats)
#, url(r'^scraping$' , scraping.Target.as_view() )
, ,
] ]
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment