Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.
Comment: Changed "it enables converting them into" to "it enables the conversion of analysis definitions and associated results data into"

Currently, analysis results are created and represented static PDF documents that may contain hundreds of tables. These tables are often difficult to navigate and there can be significant variability between sponsors. Generating these reports is expensive and they are typically only used once, offering limited reusability.

The current workflow of generating Large trials and studies generate many analysis results in the form of tables, figures, and written reports. Historically, a typical workflow for producing analysis results involves the end user generating the Analysis Data Model (ADaM) dataset, followed by generating the display in a static format such as RTF or PDF using from the Analysis Data Model (ADaM) dataset (see Figure 1). The Analysis Results Metadata (ARM) for define.xml is retrospectively generated, which provides Define-XML (available at https://www.cdisc.org/standards/foundational/define-xml/) is then created retrospectively to provide high-level metadata documentation about metadata relating to the analysis displays and results (Figure 1); however. However, there is no formal model or structures structure to describe analysis results and associated metadata which leaves , leaving a gap in standardization.  The The current process is expensive , and time-consuming, and lacks automation and traceability, leading and leads to unnecessary variation in analysis results reporting.Image Removed 

Figure 1

...

. Example of Current Workflow

Gliffy Diagram
bordertrue
displayNamehistorical process
namehistorical process
pagePin6

The goal Our vision for the future state of analysis results reporting is a world where analysis results is that they are machine-readable, easily navigable, and highly reusable. We envision the following:The aim in creating the ARS was to provide a logical model that fully described analysis results and associated metadata to support

  • automated
  • A logical model for describing analysis and results data
  • Automated generation of machine-readable results data;
  • Improved improved navigation and reusability of analysis and results data;
  • Support for the storage, access, processing, and reproducibility of results data; and
  • Traceability traceability to the study protocol/SAP , statistical analysis plan (SAP), and to the input ADaM data
  • Open-source tools for designing, specifying, building, and generating analysis results data
  • dataset.

The ARS Model has several possible implementations, including leveraging analysis results metadata to aid in automation as well as representing analysis results as data in a dataset structure. The creation of an ARS technical specification could be used to support automationTo achieve these goals, a logical model to fully describe analysis results metadata has been developed. This logical model will enable the implementation of an Analysis Results Metadata Technical Specification and an Analysis Results Data framework. ARM-TS can be used to support automation, traceability, and the creation of data displays while the ARD framework will . An analysis results dataset could support reuse and reproducibility of results data. To address the current limitations of analysis results and associated metadata reporting, we are proposing a new Figure 2 is an example of how the ARS Model could be used in a modernized workflow that shifts the focus from retrospective reporting to prospective planning.

Figure 2: Example of Potential Future Workflow

Gliffy Diagram
bordertrue
displayNameFuture Process
nameFuture Process
pagePin3

First, an end user could use the ARS Model to guide the generation of a technical specification Specifically, we propose that end-users generate the a technical specification to l prior to generating a display, rather than after the display has been created. This approach will allow allows for better planning and standardization of the analysis process, resulting in more consistent and traceable reporting. The proposed workflow (Figure 2) involves several steps. First, the end-user will develop a technical specification, which will technical specification could include metadata about the statistical methods, data sources, and displays to be generated (Figure 2, Use Case 1). Once the technical specification has been developed, the end-user will use it . This technical specification could then be used to generate an analysis results dataset , which will contain containing the results data needed to generate the display (Figure 2, Use Case 2). The analysis results dataset will could be designed to support reuse and reproducibility of the results data, enabling more efficient and effective analysis reporting.

Finally, the machine-readable analysis results dataset serves as the ‘single "single source of truth’ truth" capturing the analysis results metadata and results data in a standardized format. This dataset can then be used to generate displays for multiple reporting purposes, such as traditional analysis reporting for the clinical study report (CSR), in-text tables for the CSR, safety reporting, meta-analyses, dynamic applications, ClinicalTrials.gov, publications, and presentations. This streamlined approach ensures consistency and accuracy in the generation of displays across various deliverables, making it more efficient and reliable for reporting and communication of analysis results. 

Image Removed

Figure 2: Workflow with Future Extensions and Use Cases

Overall, this new workflow will enable allows end - users to generate analysis results metadata prospectively, with greater standardization, consistency, and traceability of analysis results reporting, enabling which in turn enables better decision - making and regulatory submissions. By shifting Shifting the focus from retrospective reporting to prospective planning , we believe that this approach will help helps to address many of the current limitations of analysis results reporting and support supports the development of more efficient and effective analysis standards.

In our vision for the future state, we anticipate the availability of open-source or community tools, such as the TFL Designer Community [1], in the industry. These tools will empower users to create machine-readable analysis metadata, which can automate the generation of analysis results data and displays. We also hope that such a tool can seamlessly integrate with existing analysis programs and report creation tools, enabling an end-to-end automation of the analysis and reporting process.

Use of LinkML for the

...

Development of the ARS Logical Model

Many of the same metadata components are needed both to create a prospective technical specification of analyses to be performed and to give context to reported results.  Therefore Therefore, a single logical model comprising has been developed, which includes the components needed to specify analyses, to represent contextualized results, and to indicate how the results are displayed.  The logical model is being developed using LinkML.

LinkML is an open-source schema development language and framework for generating machine-readable models [2]. The LinkML Generator framework generates and downstream artifacts, including JSON-Schema, ShEx, RDF, OWL, GraphQL, and SQL DDL.

The logical model acts as a blueprint or a set of rules that describes describing how things should be organized and structured. In our caseFor the ARS, it is being used to create a model that describes how analysis results data should be organized and structured. LinkML has allowed allows the creation of a standardized and consistent approach to describing analysis results data, enabling greater interoperability and collaboration across different stakeholders. Additionally, it has enables the added benefit of being able to convert them conversion of analysis definitions and associated results data into various machine-readable formats such as (e.g., JSON, YAML, OWL, or XML).

FurthermoreFurther, LinkML is a flexible and extensible tool, meaning that ; the model can be easily modified as needed to incorporate new data elements or requirements. By creating a machine-readable model, analysis results data can be more easily understood, shared, and reused by both humans and machines. LinkML will also support supports the development of validation rules that can be used to ensure the integrity and quality of the data, which  which is essential in our the highly regulated pharmaceutical industry.

...