Requirements for Re-Platforming: Difference between revisions

From CCMDB Wiki
Jump to navigation Jump to search
JMojica (talk | contribs)
No edit summary
No edit summary
 
(14 intermediate revisions by 2 users not shown)
Line 9: Line 9:


== Relatively Hard Facts ==
== Relatively Hard Facts ==
* 15-20 users on laptops, sometimes working from home, sometimes not connected to the network due to lack if [[wifi]]
* 15-20 users on laptops, sometimes working from home, sometimes not connected to the network due to lack of [[wifi]]
* About 2-2.5GB of data altogether as stored in various MS Access DBs (size may vary on other platforms)
* About 2-2.5GB of data altogether as stored in various MS Access DBs (size may vary on other platforms)
* Has several highly customized front-ends that facilitates efficient and low-error data entry
* Uses a locally customized and optimized data entry front end with facilitation/automation; using a generic tool would slow down collection and might require additional staffing to do the same work
** facilitates data entry from a daily dump received from ADT
* Has several highly customized front-ends that facilitate efficient and low-error data entry and processing
* Data we need to store is in [[Auto Data Dictionary]]
** facilitates data entry from a daily dump received from ADT (and other intermittent dumps)
** it is currently stored in [[CCMDB Data Structure]] - this structure could be stored differently but would cause large changes
* Data we store is in [[Auto Data Dictionary]]
** it is currently stored in a relational [[CCMDB Data Structure]] - this structure could be stored differently but would cause large changes
* We have (and continuously improve) [[Data Integrity Checks]]
* We have (and continuously improve) [[Data Integrity Checks]]
* Number of fields not necessarily relevant because of [[Entity–attribute–value model of the L Tmp V2 table]]
* Implemented as a system of intermittently linked MS Access databases with a fair bit of batch file and other automation facilitating their use and maintenance


== Requirements ==
== Requirements ==
''as also discussed in [[UM MedIT Re-platforming Meetings]], with decisions made tracked there and the "current master" located here. ''
''as also discussed in [[UM MedIT Re-platforming Meetings]], with decisions made tracked there and the "current master" located here. ''


Here is that draft, including for each item, a preliminary notation of its priority, on a scale of 1=lowest priority to 5=highest priority.  After the draft is completed, we will need to complete the MedIT Project Intake Process form, and then Kiran and her team will consider the best options for our needs.
Here is a draft, including for each item, a preliminary notation of its priority, on a scale of 1=lowest priority to 5=highest priority.  After the draft is completed, we will need to complete the MedIT Project Intake Process form, and then Kiran and her team will consider the best options for our needs.


'''This defines the functionality needed; the tools are up for discussion unless noted. '''


*'''Regarding data collection and uploading:'''
=== Data collection ===
**Maintaining facilitated input of admissions from the daily Shared Health data export (5)
* This would replace our current [[CCMDB.accdb]] Access front-end and would need to:
**Maintaining, without change (or with minimal change) the current data input by data collectors on the laptops  (4)
** Provide facilitated, partly automated input of admissions from the daily Shared Health data export (5)
**Allowing data collectors to add and update the data they are working on -- up until they mark a record as "complete" (5).  Currently we do this by "[[sending]]" -- although this could possible occur via some intermediary program
** Maintain a user interface that has the general look and feel of the current one (3)
**Allowing data collectors to collect chart data for patients in sites that may have poor or non-existent [[wifi]] (3) -- this relates to [[sending]] (uploading) data to the database
** Allow modification of individual data items in individual records (5)
{{DL | difficult to say, right now we have wifi access everywhere, used to be sketchy at the death desk  but we no longer are able to review charts there, SBGH does not have wifi in the basement where med records is located would we be able to use the database offline? like we currently do  [[User:Lkaita|Lisa Kaita]] 20:59, 23 March 2025 (CDT)}}
** Allow modification of multiple records via queries, programming and/or automation (5)
** Needs to work with poor or non-existent [[wifi]] (5)
** Team maintainable real-time [[Data Integrity Checks]] ([[Cross Check Engine]] implementation could provide team maintainability)


=== Data control and possibly transfer ===
* We have a "[[sending]]" process which currently includes both the movement of data from the locally-installed database to the central one, and the setting of the [[RecordStatus]] field that encodes whether the collector maintains "control" of the record or that control has been handed off to [[#Data processing]]; the collector maintains control of "incomplete" records until they "complete" the record, which triggers some mandatory [[cross check]]s for both final and [[Minimal Data Set]] data that will prevent completion unless passed.
* A new platform would need to provide this functionality:
** Make incomplete and complete data available to the [[#Data processing]] and [[#Data analysis]] stages(5)
** Allow data collectors to add and update the data they are working on -- up until they mark a record as "complete" (5)
** No longer allow them to update or view the record once it is set to "complete", or to the later stages it is set to during [[#Data processing]](5)
** If collection happens in a separate database, move and synch updated data from the [[#Data collection]] tool(5)


*'''Regarding data processing after uploading:'''
=== Data processing ===
**Allowing our database personnel to transfer data into and out of the database, such as an export to file or import from file, ad hoc (5)
* This would replace our current [[CFE]] Access front-end and would need to
**To the maximum extent possible, maintaining an interface with the data that looks like, and has functionalities of, our current MS ACCESS interface (4)
** Maintain the current [[Data Processing]] functionality for the [[data processor]], who often works remotely (5)
*** Has the ability to browse, search, sort, filter, create new variable  and export the data  from the  new platform itself (4)  --[[User:JMojica|JMojica]] 13:35, 27 March 2025 (CDT) 
** Maintain ability to run the integrity checks performed during [[Centralized data Vetting Process]], leading to the [[RecordStatus]] field being set to "vetted" if passed (5)
{{Discuss | I think this can go, it's addressed above. }}
** Maintain the ability to browse, search, sort, filter and update (add, update, delete) the data interactively, including data validation (5)
**Maintaining the current [[Data Processing]] functionality for the [[data processor]], who often works remotely (5)
** Allow modification of individual data items in individual records (5)
**Allowing the [[Statistician]] to interface with the data via SAS or additional tools (5) -- although this could possible occur via some intermediary program
** Allow modification of multiple records via queries, programming and/or automation (5)
{{DJ|
** Maintain a user interface that has the general look and feel of the current one (3)
* I don't think you write from SAS, right? [[User:Ttenbergen|Ttenbergen]] 14:33, 17 March 2025 (CDT)
** I export the ACCESS table using SAS and then perform all the analysis by programing  in SAS.  --[[User:JMojica|JMojica]] 13:35, 27 March 2025 (CDT)
* I made this a (3) because the ability to export above would provide a work-around of copying data to local; open to discuss. [[User:Ttenbergen|Ttenbergen]] 14:33, 17 March 2025 (CDT)
}}
**Allow modification individual data items in individual records (5)
**Allow modification of multiple (and possible ALL) records in a "batch" via queries and/or automation (5)
**Continued ability for us to perform [[Data Integrity Checks]], for both complete and incomplete records (5) -- currently done via ACCESS and SAS


=== Data analysis ===
* Allowing our database personnel to transfer data into and out of the database, such as an export to file or import from file, ad hoc (5)
* Retain the ability to analyze the data using other tools, including but not limited to SAS (5)


*'''Miscellaneous items:'''
=== Ongoing Improvements ===
**Maintaining the back end data format/structure as much as possible (3)
* The ability to do the following:  
**Ability, in future, to expand the capabilities of the databases by linking the data to other data obtained automatically -- e.g. Canadian Blood Services data about blood transfusions (4)
** add / remove / change fields and tables in the data structure (5)
{{Discuss | I think we would have this covered technically with the CCMDB team's ability to edit data, queries, automation and front end, so this one needs to cover the governance portion of our ability (permission?) to do this. How do we need to paraphrase it? [[User:Ttenbergen|Ttenbergen]] 14:33, 17 March 2025 (CDT)}}
** update the user interfaces to incorporate these data changes (5)
**Ability for our team to built the ETL to manage these updates, rather than rely on other teams (4)
** add / remove / change data validation and [[cross check]]s (5)
**Technical ability, in future, to link our database with the Shared Health Datamart (or name-of-the-day) (?)
* For our database personnel to do these changes independently following a reasonable change management process (4)
{{Discuss |
* What would be the priority for this?
* Technical because governance and permissions are a separate issue we would need to address; not SH's, but our service providers for any solution.
* Let's talk about this; this is where SH data will live, and if we can't interact with it in the future the utility of our data will be greatly reduced. [[User:Ttenbergen|Ttenbergen]] 14:33, 17 March 2025 (CDT) }}


* Ability for the following two things is necessarily interconnected, but listed separately to allow for some flexibility
=== Miscellaneous items ===
** Ability ''for the CCMDB team'' to update that functionality to data collection and data processing by adding to or updating front end, queries and automation, without gatekeeping
* Maintaining the back end data format/structure as much as possible (3)
{{Discuss | let's talk about this... is that a 5? It is to me.
* Maintaining our various "Created_*" queries/generated data functionality that provides APACHE score and individual element score, Charlson Score, etc
* Are all our existing queries will have a counterpart in the new platform? e.g. APACHE score and individual element score, Charlson Score, created variables, etc. --[[User:JMojica|JMojica]] 13:35, 27 March 2025 (CDT) }}
* Ability to built the ETL to manage these updates, rather than rely on other teams (4)
** Ability ''for the CCMDB team'' to modify the data structure (5)
* Ability, in future, to expand the capabilities of the databases by linking the data to other data obtained automatically -- e.g. Canadian Blood Services data about blood transfusions (4)
{{Discuss | let's talk about this... is that a 5? It is to me.}}


== Current Implementation ==
* Ability for our team to do the update and maintenance on data structure and interface (4)
* system of intermittently linked MS Access databases with a fair bit of automation
* facilitated interfaces with Shared Health systems (we request and receive data in files that we import as either temporary or permanent data)
* number of fields not necessarily relevant because of [[Entity–attribute–value model of the L Tmp V2 table]]


== Related articles ==  
== Related articles ==  
Line 75: Line 76:




[[Category: Re-platforming]]
[[Category:Re-platforming]]

Latest revision as of 21:10, 19 August 2025

This is an index for content relevant to the U of M IT team in proposing a solution for hosting our database.

Repeated failure points we should address before going too far

We have had many failed attempts at re-platforming. The technical change would be tedious but doable, the sticking points have always been around the following, so we should discuss these before planning too much further.

  • governance of any implemented system
  • data ownership
  • support model
  • our team's continued ability to change this as needed

Relatively Hard Facts

  • 15-20 users on laptops, sometimes working from home, sometimes not connected to the network due to lack of wifi
  • About 2-2.5GB of data altogether as stored in various MS Access DBs (size may vary on other platforms)
  • Uses a locally customized and optimized data entry front end with facilitation/automation; using a generic tool would slow down collection and might require additional staffing to do the same work
  • Has several highly customized front-ends that facilitate efficient and low-error data entry and processing
    • facilitates data entry from a daily dump received from ADT (and other intermittent dumps)
  • Data we store is in Auto Data Dictionary
    • it is currently stored in a relational CCMDB Data Structure - this structure could be stored differently but would cause large changes
  • We have (and continuously improve) Data Integrity Checks
  • Number of fields not necessarily relevant because of Entity–attribute–value model of the L Tmp V2 table
  • Implemented as a system of intermittently linked MS Access databases with a fair bit of batch file and other automation facilitating their use and maintenance

Requirements

as also discussed in UM MedIT Re-platforming Meetings, with decisions made tracked there and the "current master" located here.

Here is a draft, including for each item, a preliminary notation of its priority, on a scale of 1=lowest priority to 5=highest priority. After the draft is completed, we will need to complete the MedIT Project Intake Process form, and then Kiran and her team will consider the best options for our needs.

This defines the functionality needed; the tools are up for discussion unless noted.

Data collection

  • This would replace our current CCMDB.accdb Access front-end and would need to:
    • Provide facilitated, partly automated input of admissions from the daily Shared Health data export (5)
    • Maintain a user interface that has the general look and feel of the current one (3)
    • Allow modification of individual data items in individual records (5)
    • Allow modification of multiple records via queries, programming and/or automation (5)
    • Needs to work with poor or non-existent wifi (5)
    • Team maintainable real-time Data Integrity Checks (Cross Check Engine implementation could provide team maintainability)

Data control and possibly transfer

  • We have a "sending" process which currently includes both the movement of data from the locally-installed database to the central one, and the setting of the RecordStatus field that encodes whether the collector maintains "control" of the record or that control has been handed off to #Data processing; the collector maintains control of "incomplete" records until they "complete" the record, which triggers some mandatory cross checks for both final and Minimal Data Set data that will prevent completion unless passed.
  • A new platform would need to provide this functionality:
    • Make incomplete and complete data available to the #Data processing and #Data analysis stages(5)
    • Allow data collectors to add and update the data they are working on -- up until they mark a record as "complete" (5)
    • No longer allow them to update or view the record once it is set to "complete", or to the later stages it is set to during #Data processing(5)
    • If collection happens in a separate database, move and synch updated data from the #Data collection tool(5)

Data processing

  • This would replace our current CFE Access front-end and would need to
    • Maintain the current Data Processing functionality for the data processor, who often works remotely (5)
    • Maintain ability to run the integrity checks performed during Centralized data Vetting Process, leading to the RecordStatus field being set to "vetted" if passed (5)
    • Maintain the ability to browse, search, sort, filter and update (add, update, delete) the data interactively, including data validation (5)
    • Allow modification of individual data items in individual records (5)
    • Allow modification of multiple records via queries, programming and/or automation (5)
    • Maintain a user interface that has the general look and feel of the current one (3)

Data analysis

  • Allowing our database personnel to transfer data into and out of the database, such as an export to file or import from file, ad hoc (5)
  • Retain the ability to analyze the data using other tools, including but not limited to SAS (5)

Ongoing Improvements

  • The ability to do the following:
    • add / remove / change fields and tables in the data structure (5)
    • update the user interfaces to incorporate these data changes (5)
    • add / remove / change data validation and cross checks (5)
  • For our database personnel to do these changes independently following a reasonable change management process (4)

Miscellaneous items

  • Maintaining the back end data format/structure as much as possible (3)
  • Maintaining our various "Created_*" queries/generated data functionality that provides APACHE score and individual element score, Charlson Score, etc
  • Ability to built the ETL to manage these updates, rather than rely on other teams (4)
  • Ability, in future, to expand the capabilities of the databases by linking the data to other data obtained automatically -- e.g. Canadian Blood Services data about blood transfusions (4)
  • Ability for our team to do the update and maintenance on data structure and interface (4)

Related articles

Related articles: