Reviewing Survey Results & Key Issues This presentation - - PowerPoint PPT Presentation

reviewing survey results key issues
SMART_READER_LITE
LIVE PREVIEW

Reviewing Survey Results & Key Issues This presentation - - PowerPoint PPT Presentation

Reviewing Survey Results & Key Issues This presentation summarizes the key results and themes from the survey. We wish to start with a shared baseline and agreement of: Findings that may be important, but that we agree not to address at


slide-1
SLIDE 1

DMAC Steering Team Meeting | 2‐3 December 2009 | Topic: Survey Summary

Reviewing Survey Results & Key Issues

  • This presentation summarizes the key results and themes

from the survey. We wish to start with a shared baseline and agreement of: – Findings that may be important, but that we agree not to address at this meeting – Issues that need further discussion at this meeting

  • Agreeing to 3‐4 issues for discussion will help us both focus

and manage our time effectively.

1

slide-2
SLIDE 2

DMAC Steering Team Meeting | 2‐3 December 2009 | Topic: Survey Summary

2

Summary of Key Findings

  • 3.1 DMAC Purpose: More than 70% selected:

– Guide the development of the DMAC component of IOOS – Serve as a forum for issues regarding the DMAC component of IOOS – Review and comment on the progress of the DMAC component of IOOS

  • 3.2 Primary Customers:

– 66% selected IOOS related federal agencies, IOOS national office, and regional associations; 34% selected the IWGOO or its replacement – Proposed Question to Discuss: Who are the primary customers, what do they need, and what can DMAC Steering Committee provide?

slide-3
SLIDE 3

DMAC Steering Team Meeting | 2‐3 December 2009 | Topic: Survey Summary

Summary of Key Findings

  • 4.1 Content: Which of the following components of the IOOS

legislation are appropriate for DMAC team to address?

– 89% selected “Establish protocols and standards for System data processing, management, and communication” – 48% selected “develop contract certification standards and compliance procedures for all non‐Federal assets…” – 37% selected “Establish required observation data variables to be gathered by both Federal and non‐Federal assets ...”

3

slide-4
SLIDE 4

DMAC Steering Team Meeting | 2‐3 December 2009 | Topic: Survey Summary

Key Findings: Proposed Top Priorities (4.2)

  • Data Related
  • Standards Related
  • Interoperability/Infrastructure
  • Collaboration
  • DMAC Organization Related
  • Others

4

slide-5
SLIDE 5

DMAC Steering Team Meeting | 2‐3 December 2009 | Topic: Survey Summary

Summary of Key Findings

  • 4.3 Should a DMAC Steering Team continue managing a

DMAC standards adoption process? 76% “yes”; 23% “no”

  • 4.4 If DMAC Steering Team continues current standards

adoption process, what groups should be collaborated with?

– More than 70% selected NASA Standards and Process Group, Open Geospatial Consortium, International Organization for Standardization, and Global Earth Observations System of Systems – 54% selected “World Meteorological Organization”

  • 5.1 Structure: Is the current structure of the DMAC Steering

Team adequate for meeting the overall goals and objective?

– 55% of respondents (26/47) selected “no.” (44% “yes”)

5

slide-6
SLIDE 6

DMAC Steering Team Meeting | 2‐3 December 2009 | Topic: Survey Summary

Summary of Key Findings

  • 5.2 Structure: What changes, if any, would you suggest to

the current structure of the DMAC steering team?

– New/added membership composition, implementation and non‐ federal beneficiaries, international advisors/stakeholders, core team with input from regions, users, establish liaisons with standard development organizations, industry, biological side – Focus mission, Includes caucuses and expert teams, clarifying role, focus membership, single expert team, smaller team, – Keep same, re‐engage expert teams; reinforce review, monitor, and provide guidance, setting priorities, arrange around the core elements

  • f the DMAC Plan (discovery, online browse, archive)

– Funding, under‐resourced – Improve web‐presence

6

slide-7
SLIDE 7

DMAC Steering Team Meeting | 2‐3 December 2009 | Topic: Survey Summary

Summary of Key Findings

  • 5.3 Guidance: How important is it to receive regular

guidance from the IWGOO or the Interagency Ocean Observing Committee? 83%: “very important” or “important”

  • 5.4 Meetings: How often should DMAC Steering Team

meet? 43% selected “quarterly”; 31% selected “twice a year”

  • 5.5 Meetings: What is the most important outcome of a

DMAC Steering Team meeting?

– a. Guidance to DMAC Community – b. Review Standards – c. Adopt/Recommend Standards – d. Provide Updates on Related Efforts – e. Provide Feedback to Developers – f. Monitoring Goals

7

slide-8
SLIDE 8

DMAC Steering Team Meeting | 2‐3 December 2009 | Topic: Survey Summary

Common Themes Across Questions

  • Role of DMAC as a “coordinator” versus “driver” versus

“reviewer” (relates to perceived authority and scope)

  • Desired emphasis on “process” (coordination) or “product”

(outputs) (relates to overarching goal)

  • Goals with respect to external outreach and collaboration:

desired range of “sphere of influence”

  • Domain involvement versus primarily technical focus
  • Steering Team scope possibilities and limitations based on

volunteer status of many members

  • Need to establish the range of issues for DMAC Steering

Committee to take on: Breadth or Depth?

8

slide-9
SLIDE 9

DMAC Steering Team Meeting | 2‐3 December 2009 | Topic: Survey Summary

Possible Issues for Meeting Discussion

  • Defining Primary Customer(s) (3.2)

– Who are they, what do they need, and what can DMAC ST provide?

  • Top Issues to Address (4.2) and Outcomes to be Achieved (5.5)

– Given primary customers, what issues can the DMAC ST help resolve or address, and what outcomes can it best aim for?

  • Role of DMAC in Standards Adoption (4.3, 4.4) and with respect

to IOOS Legislation (4.1)

– What should these activities “look like” for DMAC Steering Committee (process and outputs)?

9

slide-10
SLIDE 10

DMAC Steering Team Meeting | 2‐3 December 2009 | Topic: Survey Summary

Issues Proposed for Deferral (1 of 2)

  • Given agreement in the survey, can we accept the following

as a starting baseline for the Purpose of DMAC (3.1), and then recheck after other issues are discussed?

– Guide the development of the DMAC component of IOOS – Forum for issues discussion regarding the DMAC component of IOOS – Review and comment on the progress of the DMAC component of IOOS

  • Identifying specific requests for guidance from IWGOO

10

slide-11
SLIDE 11

DMAC Steering Team Meeting | 2‐3 December 2009 | Topic: Survey Summary

Issues Proposed for Deferral (2 of 2)

  • Operational Considerations (5.1, 5.2, 5.3, 5.4)

– How should the DMAC ST best get guidance? Given customers, issues and desired outcomes, and roles: what should the structure be? – Note: May best be proposed after other items are considered.

  • Identifying specific organizations to collaborate with, and the

process for engaging or continuing these collaborations

  • Identifying specific proposed changes to the DMAC Steering

Team membership and a specific meeting frequency

11

slide-12
SLIDE 12

DMAC Steering Team Meeting | 2‐3 December 2009 | Topic: Survey Summary

Key Findings: Proposed Top Priorities (4.2)

  • Data Related: Establishment of operational regional and local centers for quality controlled data; Identify federal

data holdings of interest to multiple agencies or outside users; Registry; Move data from providers to users; Aiming IOOS funds at data integration; Including and integrating diverse regional observation data

  • Standards‐Related: Integrated data protocols and standards; Develop standards and protocols; Reference

libraries; Reconcile US in GOOS standards; Better integration of biological standards; Evaluate resources needed to support standards; Ensure open standards are kept open and actively used

  • Interoperability/ Infrastructure: Identification of interoperability arrangements based on international

standards for communications; Enhance coordination and interoperability between all entities; Selection of best practices in information architecture for interoperable systems; Common standards for interoperability; Advancing inter and intra agency interoperability; Ensure IOOS is compatible with international observing systems; Cyber secure infrastructure

  • Collaboration: Between agencies and related data/cyber infrastructure programs; with OOI‐CI; Between
  • bservational and modeling communities; Bridging data collectors, data servers, and data users; Coordination

with other observing systems; Define governance between agencies – who’s responsible for what

  • DMAC Organization Related: Independent assessment of DMAC Plan status; DMAC ST’s role; Sustainable DMAC;

DMAC and IOOS performance/outcome metrics; Update DMAC Data Management Plan; Stability concerning management of DMAC community;

  • Others: A design that guides activities and expenditures by National IOOS and RA’s; Meeting needs of science

community; Recommending priorities for NOAA IOOS implementation/development; Advisory role to NOAA IOOS system engineering efforts; Cultivating new technical solutions where current standards are not sufficient (funding opportunities and pilot projects); Output products; Identify funds to support joint inter‐agency technology evaluations and report lessons learned 12