scholarly journals THE METHODOLOGICAL APPROACH IN THE MANAGEMENT OF CONSERVATION DATA FOR STUCCO ARTIFACTS

2019 ◽  
pp. 65-72
Author(s):  
Filippo Edoardo Capasso ◽  
Francesca Castiello ◽  
Simona Dichiara ◽  
Manuel Giandomenico ◽  
Natalie Iacopino ◽  
...  

Some 90 Roman stucco fragments, coming from the Archaeological Park of Herculaneum, underwent in the conservation-restoration process within the Istituto Superiore per la Conservazione ed il Restauro (ISCR) laboratories in Matera. Before the conservation activity, the wide number of this manufacts brought us to design a data management system, suitable for stucco findings. Nowadays, the international approaches towards data management of this kind of artifacts is extremely heterogeneous and the absence of an unequivocal data sheet makes the cataloguing very complex and laborious. Our studies on the fragments allowed us to draw up a specific cataloguing instrument for this kind of material, carried out with a software structuring data collection in order to respond to different necessities, such as completeness and updatability, required by an efficient data management system.

2010 ◽  
Vol 18 (3) ◽  
pp. 762-766 ◽  
Author(s):  
N. Gopalakrishna Iyer ◽  
Iain J. Nixon ◽  
Frank Palmer ◽  
Ian Ganly ◽  
Snehal G. Patel ◽  
...  

CNS Spectrums ◽  
2018 ◽  
Vol 23 (1) ◽  
pp. 98-98
Author(s):  
Miriam Mina ◽  
David G. Brock ◽  
W. Scott West ◽  
Todd Hutton ◽  
Kenneth P. Pages ◽  
...  

AbstractThe NeuroStar Outcomes RegistryObjectiveNeuroStar transcranial magnetic stimulation (TMS) is an effective acute treatment for patients with major depressive disorder (MDD). In order to further understand use of the NeuroStar in a clinical setting, Neuronetics has established a patient treatment and outcomes registry to collect and analyze utilization information on patients receiving treatment with the NeuroStar.MethodsIndividual NeuroStar providers are invited to participate in the registry and agree to provide their de-identified patient treatment data. The NeuroStar has an integrated electronic data management system (TrakStar) which allows for the data collection to be automated. The data collected for the registry include Demographic Elements (age, gender), Treatment Parameters, and Clinical Ratings. Clinical assessments are: Clinician Global Impression - Severity of Illness (CGI-S) and thePatient Health Questionnaire 9-item (PHQ-9). De-identified patient data is uploaded to Registry server; an independent statistical service then creates final data reports.ResultsOver 500 patients have entered the NeuroStar Outcomes Registry since Sept 2016. Mean patient age: 48.0 (SD±16.0); 64% Female. Baseline PHQ-9, mean 18.8 (SD±5.0.) Response/Remission Rate, PHQ-9: 61%/33% CGI-S: 78%/59%.ConclusionsFor the initial 500 patients in the Outcomes Registry, approximately 2/3 patients achieve respond and 1/3 patients achieve remission with an acute course of NeuroStar. These treatment outcomes consistent with NeuroStar open-label study data (Carpenter, 2012). The TrakStar data management system makes large scale data collection feasible. The NeuroStarOutcomes Registry is ongoing, and expected to reach 6000 outpatients from more than 47 clinical sites in 36 months.Funding AcknowledgementsNeuronetics, Inc.


2016 ◽  
Vol 131 (2_suppl) ◽  
pp. 41-43
Author(s):  
Andrew J. Hebert ◽  
Tamara L. Lamia ◽  
Ben T. Schoenbachler ◽  
Ayasha K. Richardson

2017 ◽  
Vol 143 ◽  
pp. 173-178 ◽  
Author(s):  
Arun Kumar ◽  
Pushpendu Kar ◽  
Rakesh Warrier ◽  
Aditi Kajale ◽  
Sanjib Kumar Panda

2017 ◽  
Vol 4 (1) ◽  
pp. 62-66
Author(s):  
Luyen Ha Nam

From long, long time ago until nowadays information still takes a serious position for all aspect of life, fromindividual to organization. In ABC company information is somewhat very sensitive, very important. But how wekeep our information safe, well we have many ways to do that: in hard drive, removable disc etc. with otherorganizations they even have data centre to save their information. The objective of information security is to keep information safe from unwanted access. We applied Risk Mitigation Action framework on our data management system and after several months we have a result far better than before we use it: information more secure, quickly detect incidents, improve internal and external collaboration etc.


Shore & Beach ◽  
2020 ◽  
pp. 92-101
Author(s):  
Richard Raynie ◽  
Syed Khalil ◽  
Charles Villarrubia ◽  
Ed Haywood

The Coastal Protection and Restoration Authority (CPRA) of Louisiana was created after the devastating hurricanes of 2005 (Katrina and Rita) and is responsible for planning and implementing projects that will either reduce storm-induced losses (protection) or restore coastal ecosystems that have been lost or are in danger of being lost (restoration). The first task of the CPRA board was to develop Louisiana’s first Coastal Master Plan (CPRA 2007), which formally integrates and guides the protection and restoration of Louisiana’s coast. The System-Wide Assessment and Monitoring Program (SWAMP) was subsequently developed as a long-term monitoring program to ensure that a comprehensive network of coastal data collection activities is in place to support the planning, development, implementation, and adaptive management of the protection and restoration program and projects within coastal Louisiana. SWAMP includes both natural-system and human-system components and also incorporates the previously-developed Coastwide Reference Monitoring System (CRMS), the Barrier Island Comprehensive Monitoring (BICM) program, and fisheries data collected by the Louisiana Department of Wildlife and Fisheries (LDWF) in addition to other aspects of system dynamics, including offshore and inland water-body boundary conditions, water quality, risk status, and protection performance, which have historically not been the subject of CPRA-coordinated monitoring. This program further facilitates the integration of project-specific data needs into a larger, system-level design framework. Monitoring and operation of restoration and protection projects will be nested within a larger hydrologic basin-wide and coast-wide SWAMP framework and will allow informed decisions to be made with an understanding of system conditions and dynamics at multiple scales. This paper also provides an update on the implementation of various components of SWAMP in Coastal Louisiana, which began as a Barataria Basin pilot implementation program in 2015. During 2017, the second phase of SWAMP was initiated in the areas east of the Mississippi River. In 2019, development of SWAMP design was completed for the remaining basins in coastal Louisiana west of Bayou Lafourche (Figure 1). Data collection is important to inform decisions, however if the data are not properly managed or are not discoverable, they are of limited use. CPRA is committed to ensuring that information is organized and publicly available to help all coastal stakeholders make informed, science-based decisions. As a part of this effort, CPRA has re-engineered its data management system to include spatial viewers, tabular download web pages, and a library/document retrieval system along with a suite of public-facing web services providing programmatic access. This system is collectively called the Coastal Information Management System (CIMS). CPRA and U.S. Geological Survey (USGS) are also developing a proposal to create an interface for CIMS data to be exported to a neutral template that could then be ingested into NOAA’s Data Integration Visualization, Exploration and Reporting (DIVER) repository, and vice versa. DIVER is the repository that the Natural Resource Damage Assessment (NRDA) program is using to manage NRDA-funded project data throughout the Gulf of Mexico. Linking CIMS and DIVER will make it easier to aggregate data across Gulf states and look at larger, ecosystem-level changes.


Sign in / Sign up

Export Citation Format

Share Document