Call information:
https://uqam.zoom.us/j/85478694266
Meeting Attendees
- Héon, Michel
- Benjamin Kampe
- Sandra Mierz
- Graham Triggs
- Andrew Woods
- Belkouch, Rachid
- Dickner, Nicolas
Sprint vision:
- Connecting VIVO to ORCID by Kafka
Discussions
- Triggs : plan to use Kafka, but not for ingestion ; would be interested to see what Kafka can do in this context, though
- Kampe: big problem shouldn't be to integrate from Kafka to VIVO, but rather out of VIVO to Kafka
- Triggs: our interest would be to see how to trigger events from VIVO ; (used it to trigger modeling, indexing, etc) ; rather straight forward, but to push a message out to Kafka, in order to distribute it to other systems should be trickier
- Woods: mechanics is already being worked on, but have to figure out the nature of the body
- Héon: so messaging mechanism per se isn't available yet in VIVO? (Andrew: yes)
- Héon: so building an extractor from VIVO would be a goal (Kampe: yes)
- Héon: logistics ; can we use Lyrasis' wiki, GitHub, Slack, etc?
- Woods : absolutely ; though creating a new channel might be more appropriate that using the data ingestion channel
- Woods : re: Slack channel, doesn't think that the TIB use case is really data ingest,
- Belkouch: agreed, but would like to foster interest to a wider base
- Triggs: we have more pressing needs for data exchange than for data ingest
- Mierz: has a projet of her own with U of Osnabrück, so she'd be very interested in data ingest
- Belkouch: does TIB has membership access to ORCID APIs? (Triggs: doesn't think so)
- Héon: important to use the bidirectionnal API ; possible public sandbox
- Woods : yes, there should be such a sandbox + we could reach out to the US consortium ; also: folks from the Smithsonians have been working on a similar project
- Belkouch: any useful known resources about ORCID API?
- Triggs: there is code to handle connection to ORCID ; not sure how useful it would be in the circumstances
- Woods : considering use case of TIB (Kafka as a data consumer), would it make to to coordinate with William Welling effort?
- Triggs: pressing need, no time to coordinate with several projects ; aiming at delivering
- Both agree to see whether they can collaborate ; Rachid will organize short meeting next week
- Héon to Woods : methodological advices?
- Woods : seems like there are two tracks to the sprint: data ingest (UQAM + Mierz) and data exchange (Kampe+Triggs), and important not to slow down people by calling them into meetings with too wide scopes(?)
- Andrew will be busy with other sprint next week.
- More logistic discussions.
- Andrew will 1) see to creating of a space on Wiki, 2) tell about it to relevant people in Lyrasis, 3) get in touch with his contact in Smithsonian'
- Michel will create standup page on wiki
Sprint objectives
- Exploring the use of Kafka to connect ORCID
- Validate accesses to ORCID
- Identifying ORCID APIs that may be useful
- Architectural plan for the deployment of Kafka
- Effort/benefit analysis in the use of Kafka
Criteria for success
- At the end of the sprint, having in hand the necessary elements for a go/no-go decision on the opportunity to use Kafka as a VIVO data ingest mechanism.
Communication channels
- Wiki UQAM-VIVO-PUBLIC
- Slack Chanal of Lyrasys ?
- Git Repo Where ?
References
- VIVO-Conf 2020: VIVO-DATACONNECT: TOWARDS AN ARCHITECTURAL MODEL FOR INTERCONNECTING HETEROGENEOUS DATA SOURCES TO POPULATE THE VIVO TRIPLESTORE
- Kafka Home Page
- Introduction - Apache Kafka - The Apache Software Foundation!