{-| Module : Gargantext.API.Node.DocumentsFromWriteNodes Description : Copyright : (c) CNRS, 2017-Present License : AGPL + CECILL v3 Maintainer : team@gargantext.org Stability : experimental Portability : POSIX -} {-# LANGUAGE MonoLocalBinds #-} {-# LANGUAGE TemplateHaskell #-} {-# LANGUAGE TypeOperators #-} module Gargantext.API.Node.DocumentsFromWriteNodes where -- import Data.Maybe (fromMaybe) import Conduit import Control.Lens ((^.)) import Data.Aeson import Data.Either (Either(..), rights) import Data.Swagger import GHC.Generics (Generic) import Gargantext.API.Admin.EnvTypes (Env, GargJob(..)) import Gargantext.API.Admin.Orchestrator.Types (JobLog(..), AsyncJobs) import Gargantext.API.Admin.Types (HasSettings) import Gargantext.API.Job (jobLogSuccess, jobLogFailTotalWithMessage) import Gargantext.API.Prelude (GargM, GargError) import Gargantext.Core (Lang(..)) import Gargantext.Core.Text.Corpus.Parsers.FrameWrite import Gargantext.Core.Text.List.Social (FlowSocialListWith) import Gargantext.Core.Text.Terms (TermType(..)) import Gargantext.Core.Types.Individu (User(..)) import Gargantext.Database.Action.Flow (flowDataText, DataText(..)) import Gargantext.Database.Action.Flow.Types (FlowCmdM) import Gargantext.Database.Admin.Types.Hyperdata.Document (HyperdataDocument(..)) import Gargantext.Database.Admin.Types.Hyperdata.Frame import Gargantext.Database.Admin.Types.Node import Gargantext.Database.Query.Table.Node (getChildrenByType, getClosestParentIdByType', getNodeWith) import Gargantext.Database.Schema.Node (node_hyperdata, node_name, node_date) import Gargantext.Prelude import Gargantext.Utils.Jobs (serveJobsAPI) import Gargantext.Core.Text.Corpus.Parsers.Date (split') import Servant import qualified Data.List as List import qualified Data.Text as T -- import qualified Gargantext.Defaults as Defaults ------------------------------------------------------------------------ type API = Summary " Documents from Write nodes." :> AsyncJobs JobLog '[JSON] Params JobLog ------------------------------------------------------------------------ data Params = Params { id :: Int , paragraphs :: Int , lang :: Lang , selection :: FlowSocialListWith } deriving (Generic, Show) instance FromJSON Params where parseJSON = genericParseJSON defaultOptions instance ToJSON Params where toJSON = genericToJSON defaultOptions instance ToSchema Params ------------------------------------------------------------------------ api :: UserId -> NodeId -> ServerT API (GargM Env GargError) api uId nId = serveJobsAPI DocumentFromWriteNodeJob $ \p log'' -> let log' x = do liftBase $ log'' x in documentsFromWriteNodes uId nId p (liftBase . log') documentsFromWriteNodes :: (HasSettings env, FlowCmdM env err m) => UserId -> NodeId -> Params -> (JobLog -> m ()) -> m JobLog documentsFromWriteNodes uId nId Params { selection, lang, paragraphs } logStatus = do let jobLog = JobLog { _scst_succeeded = Just 1 , _scst_failed = Just 0 , _scst_remaining = Just 1 , _scst_events = Just [] } logStatus jobLog mcId <- getClosestParentIdByType' nId NodeCorpus cId <- case mcId of Just cId -> pure cId Nothing -> do let msg = T.pack $ "[G.A.N.DFWN] Node has no corpus parent: " <> show nId logStatus $ jobLogFailTotalWithMessage msg jobLog panic msg frameWriteIds <- getChildrenByType nId NodeFrameWrite -- https://write.frame.gargantext.org//download frameWrites <- mapM (\id -> getNodeWith id (Proxy :: Proxy HyperdataFrame)) frameWriteIds frameWritesWithContents <- liftBase $ mapM (\node -> do contents <- getHyperdataFrameContents (node ^. node_hyperdata) pure (node, contents) ) frameWrites let parsedE = (\(node, contents) -> hyperdataDocumentFromFrameWrite lang paragraphs (node, contents)) <$> frameWritesWithContents let parsed = List.concat $ rights parsedE _ <- flowDataText (RootId (NodeId uId)) (DataNew (Just $ fromIntegral $ length parsed, yieldMany parsed)) (Multi lang) cId (Just selection) logStatus pure $ jobLogSuccess jobLog ------------------------------------------------------------------------ hyperdataDocumentFromFrameWrite :: Lang -> Int -> (Node HyperdataFrame, T.Text) -> Either T.Text [HyperdataDocument] hyperdataDocumentFromFrameWrite lang paragraphSize (node, contents) = case parseLines contents of Left _ -> Left "Error parsing node" Right (Parsed { authors, contents = ctxts}) -> let HyperdataFrame { _hf_base, _hf_frame_id } = node ^. node_hyperdata authorJoinSingle (Author { firstName, lastName }) = T.concat [ lastName, ", ", firstName ] authors' = T.concat $ authorJoinSingle <$> authors --{- (year',month',day') = split' (node^. node_date) date' = Just $ T.concat [ T.pack $ show year', "-" , T.pack $ show month', "-" , T.pack $ show day' ] --} {- date' = (\(Date { year, month, day }) -> T.concat [ T.pack $ show year', "-" , T.pack $ show month', "-" , T.pack $ show day' ]) <$> date year' = fromIntegral $ maybe Defaults.year (\(Date { year }) -> year) date month' = maybe Defaults.month (\(Date { month }) -> fromIntegral month) date day' = maybe Defaults.day (\(Date { day }) -> fromIntegral day) date --} in Right (List.map (\(t, ctxt) -> HyperdataDocument { _hd_bdd = Just "FrameWrite" , _hd_doi = Nothing , _hd_url = Nothing , _hd_uniqId = Nothing , _hd_uniqIdBdd = Nothing , _hd_page = Nothing , _hd_title = Just t , _hd_authors = Just authors' , _hd_institutes = Nothing , _hd_source = Just $ node ^. node_name , _hd_abstract = Just ctxt , _hd_publication_date = date' , _hd_publication_year = Just year' , _hd_publication_month = Just month' , _hd_publication_day = Just day' , _hd_publication_hour = Nothing , _hd_publication_minute = Nothing , _hd_publication_second = Nothing , _hd_language_iso2 = Just $ T.pack $ show lang } ) (text2titleParagraphs paragraphSize ctxts) )