Roadmap: Global research data management advisory platform combines DMPTool and DMPonline

Roadmap, a global data management advisory platform that links data management plans (DMPs) to other components of the research lifecycle is a new open science initiative from partners at the University of California Curation Center (UC3) of the California Digital Library (CDL), USA, and the Digital Curation Centre (DCC), United Kingdom.

Both organizations sponsor and maintain such platforms, the DMPTool and DMPonline respectively. They allow researchers from around the world to create their data management plans in less time by employing ready-to-use templates with specific guidance tailored to address the requirements of specific funding agencies in the USA and the UK.

Recently, the proliferation of data sharing policies throughout the world has produced increasing demand for data management planning support from both organizations. Therefore, it makes sense for the CDL and DCC to consolidate efforts and move beyond a focus on national researchers and funders to extend their global outreach through Roadmap, a new open-source platform for data management planning. Their proposal was submitted to the Open Science Prize contest and is now published in the open access journal Research Ideas and Outcomes (RIO).

While the two teams have been working together unofficially and engaging in international initiatives, a formal partnership would signal to the global research community that there is one place to create DMPs and find advisory information.

“Research data management (RDM) that enables open science is now acknowledged as a global challenge: research is global, policies are becoming global, and thus the need is global,” explain the authors. “Open science has a global agenda, and by making DMPs true infrastructure in a global open access community we will elevate research and open data for reuse.”

roadmap still

In their joint project, the two organizations will combine their experience along with all existing functionality from their tools regarding the DMP use case into a single technical platform.

“New work on our respective systems is already underway to enable internationalization, integrate with other organizations and technical platforms, and encourage greater openness with DMPs,” they explain. “By joining forces, the Roadmap system will consolidate these efforts and move beyond a narrow focus on specific funders in specific countries, and even beyond institutional boundaries, to create a framework for engaging with disciplinary communities directly.”

To facilitate data sharing, reuse, and discoverability, Roadmap will be integrated with a number of platforms such as the Open Science Framework, SHARE, the Crossref/Datacite DOI Event Tracking system and Zenodo, among others. “Linking systems and research outputs across the web increases the chances that data will be discovered, accessed, and (re)used,” note the authors.

The team’s plan for enhanced openness includes encouraging authors to share their newly created data management plans by setting their privacy on “public” by default. They also intend to assign digital object identifiers (DOIs) to all plans, thus making them citable and motivating their authors to make them openly accessible. As part of this initiative, five researchers have just published their DMPs, created with the DMPTool, in Research Ideas and Outcomes (RIO).

“We see greater potential for the DMP as a dynamic checklist for pre- and post-award reporting; a manifest of research products that can be linked with published outputs; and a record of data, from primary through processing stages, that could be passed to repositories,” state the authors. “The DMP will therefore not only support the management of the data but boost its discoverability and reuse.”

###

Original source:

Simms S, Jones S, Ashley K, Ribeiro M, Chodacki J, Abrams S, Strong M (2016) Roadmap: A Research Data Management Advisory Platform. Research Ideas and Outcomes 2: e8649. doi: 10.3897/rio.2.e8649

Empowering stakeholders: FP7 project EU BON shares know-how on biodiversity data policies

Engagement with relevant political authorities and other stakeholders is of crucial importance for a research project, making sure its objectives are in tune with the real-world problems and its results provide adapted solutions. The EU-funded FP7 project Building the European Biodiversity Observation Network (EU BON) shares the outcomes, lessons learned and conclusions from a series of three roundtable meetings designed to identify stakeholder needs and promote collaboration between science and policy.

The collection of EU BON stakeholder roundtable reports provides a summarized overview of shared experiences gained in the three different workshops that were organized from 2013-2016. With more than 100 participants from over 20 countries altogether, the roundtable reports provide insights and exchange of ideas on highly relevant issues concerning policy, citizen science and local/regional stakeholders and its networks.

The roundtables seek to build up a stakeholder dialogue with exemplary sector-specific user communities to incorporate feedback loops for the products of EU BON, as well as to develop improvements of existing biodiversity data workflows. Being published via the innovative Research Ideas & Outcomes (RIO) journal conclusions, derived knowledge and results are now made available for other projects and the wider community to ensure their re-use.

The three roundtable papers report on conclusion on highly relevant issues related to biodiversity information and its open-access and availability, data workflows and integration of citizen science as well as science-policy interfaces.

“In each of the three detailed reports of the roundtables we outline its aims, intentions, as well as results and recommendations, that were drafted based on the roundtable discussions, world café sessions and working groups. Such project results are now published for the first time in the new series of EU BON results, featured in RIO, providing a unique new medium to share experiences, outcomes and conclusions,” comments Dr. Katrin Vohland, Museum für Naturkunde, Berlin.

“The three reports were published as workshop report provided by the Research Ideas & Outcomes (RIO) journal. This allows readers to publish, distribute and computationally analyse myriads of workshop reports that otherwise often get forgotten or just lost,” comments Prof. Lyubomir Penev, co-founder and publisher of RIO.

###

Original Sources:

Rationale of the roundtables

Wetzel F, Hoffmann A, Häuser C, Vohland K (2016) 1st EU BON Stakeholder Roundtable (Brussels, Belgium): Biodiversity and Requirements for Policy. Research Ideas and Outcomes 2: e8600. doi: 10.3897/rio.2.e8600

Vohland K, Häuser C, Regan E, Hoffmann A, Wetzel F (2016) 2nd EU BON Stakeholder Roundtable (Berlin, Germany): How can a European biodiversity network support citizen science? Research Ideas and Outcomes 2: e8616. doi: 10.3897/rio.2.e8616

Vohland K, Hoffmann A, Underwood E, Weatherdon L, Bonet F, Häuser C, Wetzel F (2016) 3rd EU BON Stakeholder Roundtable (Granada, Spain): Biodiversity data workflow from data mobilization to practice. Research Ideas and Outcomes 2: e8622. doi: 10.3897/rio.2.e8622

General synthesis and lessons learnt from the three EU BON stakeholder roundtables

###

About EU BON:

EU BON stands for “Building the European Biodiversity Observation Network” and is a European research project, financed by the 7th EU framework programme for research and development (FP7). EU BON seeks ways to better integrate biodiversity information and implement into policy and decision-making of biodiversity monitoring and management in the EU.

Making the most out of biological observations data

Creating and maintaining a biodiversity data collection has been a much-needed worldwide exercise for years, yet there is no single standard on how to do this. This has led to a myriad of datasets often incompatible with each other. To make the most out of biodiversity data and to ensure that its use for environmental monitoring and conservation is both easy and legal, the FP7-funded EU project Building the European Biodiversity Observation Network (EU BON) published recommendations that provide consistent Europe-wide Data Publishing Guidelines and Recommendations in the EU BON Biodiversity Portal.

The report “Data Policy Recommendations for Biodiversity Data. EU BON Project Report” featured in the Research Ideas & Outcomes (RIO) journal, is the first contribution in a pioneering comprehensive project outputs compilation taking advantage of RIO’s unique option to publish collections of project results.

Biodiversity data and information provide important knowledge for many biological, geological, and environmental research disciplines. Additionally, they are crucial for the development of strong environmental policies and the management of natural resources. Information management systems can bring together a wealth of information and a legacy of over 260 years of biological observations which are now dispersed in a myriad of different documents, institutions, and locations.

EU BON aims to build a comprehensive “European Biodiversity Portal” that will incorporate currently scattered Europe-wide biodiversity data, while at the same time helping to realize a substantial part of the worldwide Group on Earth Observations Biodiversity Observation Network (GEO BON). To achieve this ambitious plan, EU BON identifies the strong need for a coherent and consistent data policy in Europe to increase interoperability of data and make its re-use both easy and legal.

“Biodiversity data and information should not be treated as commercial goods, but as a common resource for the whole human society. The EU BON data sharing agreement is an important step in this direction,” comments the lead author of the report Dr. Willi Egloff from Plazi, Switzerland.

In its report, the EU BON project analysis available single recommendations and guidelines on different topics. On this basis, the report provides structured guidelines for legislators, researchers, data aggregators, funding agencies and publishers to be taken into consideration towards providing standardized, easy-to-find, re-shareable and re-usable biodiversity data.

“We are extremely happy that EU BON is among the first to take advantage of our project outputs collections option in RIO. The first report they are publishing with us deals with issues of opening up data, and digitizing and collecting scientific knowledge, all close to RIO’s mission to open up the research process and promote open science,” says Prof. Lyubomir Penev, Founder and Publisher of RIO.

###

Original Source:

Egloff W, Agosti D, Patterson D, Hoffmann A, Mietchen D, Kishor P, Penev L (2016) Data Policy Recommendations for Biodiversity Data. EU BON Project Report. Research Ideas and Outcomes2: e8458. doi: 10.3897/rio.2.e8458

 

About EU BON:

EU BON stands for “Building the European Biodiversity Observation Network” and is a European research project, financed by the 7th EU framework programme for research and development (FP7). EU BON seeks ways to better integrate biodiversity information and implement into policy and decision-making of biodiversity monitoring and management in the EU.

One place for all scholarly literature: An Open Science Prize proposal

Openly accessible scholarly literature is referred to as “the fabric and the substance of Open Science” in the present small grant proposal, submitted to the Open Science Prize contest and published in the Research Ideas and Outcomes (RIO) open access journal. However, the scholarly literature is currently quite chaotically dispersed across thousands of different websites and disconnected from its context.

To tackle this issue, authors Marcin Wojnarski, Paperity, Poland, and Debra Hanken Kurtz, DuraSpace, USA, build on the existing prototype Paperity, the first open access aggregator of scholarly journals. Their suggestion is the first global universal catalog of open access scientific literature. It is to bring together all publications by automatically harvesting both “gold” and “green” ones.

Called Paperity Central, it is to also incorporate many helpful functionalities and features, such as a wiki-type one, meant to allow for registered users to manually improve, curate and extend the catalog in a collaborative, community-controlled way.

“Manual curation will be particularly important for “green” metadata, which frequently contain missing or incorrect information; and for cataloguing those publications that are inaccessible for automatic harvesting, like the articles posted on author homepages only,” further explain the authors.

To improve on its ancestor, the planned catalog is to seamlessly add “green” publications from across repositories to the already available articles indexed from gold and hybrid journals. Paperity Central is to derive its initial part of “green” content from DSpace, the most popular repository platform worldwide, developed and stewarded by DuraSpace, and powering over 1,500 academic repositories around the world.

All items available from Paperity Central are to be assigned with globally unique permanent identifiers, thus reconnecting them to their primary source of origin. Moreover, all different types of Open Science resources related to a publication, such as author profiles, institutions, funders, grants, datasets, protocols, reviews, cited/citing works, are to be semantically linked in order to assure none of them is disconnected from its context.

Furthermore, the catalog is to perform deduplication of each entry in the same systematic and consistent way. Then, these corrections and expansions are to be transferred back to the source repositories in a feedback loop via open application programming interfaces (APIs). However, being developed from a scratch, its code will possess many distinct features setting it apart from existing wiki-type platforms, such as Wikipedia, for example.

“Every entry will consist of structured data, unlike Wikipedia pages which are basically text documents,” explain the scientists. “The catalog itself will possess internal structure, with every item being assigned to higher-level objects: journals, repositories, collections – unlike Wikipedia, where the corpus is a flat list of articles.”

In order to guarantee the correctness of the catalog, Paperity Central is to be fully transparent, meaning the history of changes is to be made public. Meanwhile, edits are to be moderated by peers, preferably journal editors or institutional repository admins overlooking the items assigned to their collections.

In their proposal, the authors note that the present development plan is only the first phase of their project. They outline the areas where the catalog is planned to be further enhanced in future. Among others, these include involvement of more repositories and platforms, fully developed custom APIs and expansion on the scholarly output types to be included in the catalog.

“If we are serious about opening up the system of scientific research, we must plant it on the foundation of open literature and make sure that this literature is properly organized and maintained: accessible for all in one central location, easily discoverable, available within its full context, annotated and semantically linked with related objects,” explain the scientists.

“Assume we want to find all articles on Zika published in 2015,” they exemplify. “We can find some of them today using services like Google Scholar or PubMed Central, but how do we know that no other exist? Or that we have not missed any important piece of literature? With the existing tools, which have incomplete and undefined coverage, we do not know and will never know for sure.”

In the spirit of their principles of openness, the authors assure that once funded, Paperity Central will be releasing its code as open source under an open license.

###

Original source:

Wojnarski M, Hanken Kurtz D (2016) Paperity Central: An Open Catalog of All Scholarly Literature. Research Ideas and Outcomes 2: e8462. doi: 10.3897/rio.2.e8462

Openly published Open Science Prize Grant Proposal builds on ContentMine and Hypothes.is to bridge scientists and facts

Public health emergencies such as the currently spreading Zika disease might be successfully necessitating open access for the available biomedical researches and their underlying data, yet filtering the right information, so that it lands in the hands of the right people, is what holds up professionals to bring the adequate measures about.

Submitted to the Open Science Prize contest, the present grant proposal, prepared with the joint efforts of scientists affiliated with Hypothes.is, ContentMine, University of CambridgeCottage Labs LLP and Imperial College of London, suggests a new scholarly assistant system, called amanuens.is, based on the existing ContentMine and Hypothes.is prototypes. Its aim is to combine machines and humans, so that mining critically important facts and making them available to the world can be made not only significantly faster, but also less costly. Through their publication in the open access journal Research Ideas and Outcomes (RIO), the scientists, who are also well-known open access and open data proponents, are looking for further support, feedback and collaborations.

While Hypothes.is is a mixture of software and communities, which together annotate the available literature, ContentMine are building an open source pipeline to extract facts from scientific documents, thus making the literature review process cheaper, more rigorous, continuous and transparent. The role of amanuens.is is meant to bring these two systems together.

As a result, Hypothes.is is to display ContentMine facts as annotations on the online document, therefore increasing their visibility. In turn, the large Hypothes.is community, comprising users ranging from devoted and experienced Wikipedia editors to dedicated citizen scientists, would be able to provide manually their own annotations, which could be then fed back into the ContentMine facts store.

“Facts are important – but science is performed by people – so ContentMine are partnering with Hypothes.is to bring communities together around facts in the scholarly literature,” sums up Dr Peter Murray-Rust. “Through combining machines and humans in a tight, iterating, loop, amanuens.is will be able to mine critically important facts and make them available to the world.”

In their proposal, the authors give a hypothetical, yet foreseeable example with a Hypothes.is community, centered around research and discussions regarding a bacterium, already proven to restrain some mosquitoes from transmitting various viruses, and its potential use against Zika. There, amanuens.is downloads all open access papers on Zika from a multitude of sources within 3 minutes. In a matter of a couple of seconds a total of 123 files are downloaded. Then, amanuens.is delivers a data table of the extracted data, including species, human genes, DNA primers and top word frequencies.

Within the community and thanks to the literature, made available via ContentMine, the users would be able to collaborate and build on the existing research outcomes. As a result, it could take only fifteen minutes and a brief proposal to mobilise the related scholarly resources and test for Zika resistance in infected with the virus mosquitoes.

“Finding facts to finding people took 15 minutes and this is how modern collaborative science should work,” Prof Peter Murray-Rust says about the given example. “The people then create knowledge from the facts. The knowledge creates communities. The communities explore science- and people-based solutions.”

In conclusion, the proposal states that similarly to the content and software provided by ContentMine and Hypothes.is, the outputs produced by amanuens.is will also be openly available. All of its data and annotations are to be public domain under a CC0 waiver.

###

Original source:

Martone M, Murray-Rust P, Molloy J, Arrow T, MacGillivray M, Kittel C, Kasberger S, Steel G, Oppenheim C, Ranganathan A, Tennant J, Udell J (2016) ContentMine/Hypothes.is Proposal.Research Ideas and Outcomes 2: e8424. doi: 10.3897/rio.2.e8424

Soilless farming suggested as a solution to food shortage in Qatar

Soilless farming could help developing countries with little arable land and harsh for agriculture climate, such as Qatar, to become self-sufficient in terms of their produce. Relying on advanced hydroponics and multi-story vertical growing, the proposed system uses nutrient-enriched water to produce approximately a hundred times more yield compared to when the crops are grown on a conventional farmland of the same size.

The hybrid setup, devised by Nik-Othman Abdullah, biotechnologist at Malaysia University of Science and Technology, is described in his Methods paper, published in the open-access peer-reviewed journal Research Ideas and Outcomes (RIO). In his publication he describes and explains the steps of the approach meant to tackle the food-shortage and enormous import expenditures in the country in line with the Ministry of Environment of Qatar’s recently created National Food Security Programme.

The proposed vertical-horizontal regulated soilless farming is theoretically capable of increasing the domestic produce on such a scale that the country, which has been estimated to have spent USD 11 billion on imported food in 2014 alone, could become self-sufficient.

Furthermore, this type of soilless farming could provide reliable quantity as well as quality of the crops. Grown indoors, where they would be constantly monitored by personnel with good technical and scientific knowledge, the produce would be less affected by factors such as atmospheric conditions, contamination or pests.

The plants would be supplied with the calculated amount of nutrition they need, as well as the exact amount of light and gas exposure. Being grown in a sterile environment and not treated with fertilizers, pesticides, weedicides and other harsh chemicals, the crops would not only look visibly identical, but would also be cleaner, fresher, healthier, tastier and richer in nutrient content. They would also grow faster and bigger.

“Plants would not waste energy in root tissue production because nutrients in pure form will be provided to the plants instead of the plant stressing to search for the nutrients,” explains the biotechnologist. “Therefore, plants grow evidently 50% faster and bigger.”

On the other hand, such farming would be more cost-efficient, since there would be significantly less personnel needed and no expenditure on chemical treatment. Moreover, it would take only 10% of the water used in conventional farming for a same-sized piece of agricultural land. Overall, soilless farming would cost 90% less, although the initial setup would be quite expensive. The latter is also the only disadvantage pointed out in the paper.

The author stresses that such a farming ground can be constructed basically in any location. “It can be set up almost anytime and everywhere, in a greenhouse, warehouse, inside a building, or even in outer space,” Nik-Othman Abdullah comments.

###

Original source:

Abdullah N (2016) Vertical-Horizontal Regulated Soilless Farming via Advanced Hydroponics for Domestic Food Production in Doha, Qatar. Research Ideas and Outcomes 2: e8134. doi:10.3897/rio.2.e8134

RIO @ OpenCon

OpenCon 2015 brought together students and early career academic professionals alongside the most passionate advocates for Open Science to talk about Open Access, Open Education, and Open Data… and RIO’s founding editors Ross Mounce and Daniel Mietchen were there too!

There is hardly a better place to meet young and devoted enthusiasts supporting openness and transparency for scientific research. This is also a good setting to talk about  the RIO concept.

Special thanks to keynote speaker and member of our Subject Editor team Erin McKiernan for her passionate pledge for opening up science and for mentioning RIO Journal in her talk the very first day!

keynote-erin-mckiernan-my-pledge-to-be-open-yeah-hows-that-going-7-638    keynote-erin-mckiernan-my-pledge-to-be-open-yeah-hows-that-going-10-638
Slides from Erin McKiernan at Open Con 2015: My Pledge to be open (Yeah, how’s that going?) http://www.slideshare.net/RightToResearch/keynote-erin-mckiernan-my-pledge-to-be-open-yeah-hows-that-going

As part of the 3-min community presentations, the journal was also presented by Daniel Mietchen, who has been a central figure and inspiration behind the RIO idea and its implementation.

Further discussions followed. Created around the principles of Open Access and Open Data, RIO aims to open up science making it possible to publish every step of the research cycle, from ideas to final results. See the full range of publications here.

RIO is devoted to link research to its real social impact. The journal will apply categorization to research outputs based upon internationally-recognised external criteria such as the EU Societal Challenges (the priorities of EU Horizon 2020 funding), and the United Nations’ Sustainable Development Goals (SDGs).

We had some fun too. RIO sent some T-shirts and stickers for our supporters to take and spread the word: RIO flyers and stickers will soon be available in Rio de Janeiro, just ask Iara!

Special thanks to everyone for the support! We look forward to more feedback from the community, please do not shy away to ping us at rio@riojournal.com or on Twitter @RioJournal.  

RIO is open for submissions since the beginning of November, see our APCs and promotions. We also welcome applications for Subject Editors and Editorial Apprentices.

RIO pricing: affordable, flexible, sustainable

Today, we are happy to announce our initial pricing scheme and launch promotions.

Many people have asked us what we are going to charge for our novel services – we welcome such questions. In an age when some are charging in excess of $5000 for publishing a single open-access article, we are proud to keep our prices low, but sustainable.

You may notice the pricing structure is more complicated than usual. Don’t Panic. In this blog post, we’ll break it down into easy pieces. We think this flexibility of pricing is a positive feature: it reflects that some outputs are less costly to publish, and others are more expensive. We also offer à la carte pricing with respect to optional services such as linguistic editing, PR services and paper-printed copies.

Try RIO for free

RIO is already open for submissions. To encourage you to give it a try, we’re making idea-stage submissions completely free from now on until the end of April 2016. This promotional offer applies to the following output types if submitted through the ARPHA writing tool: Small Grant Proposals, Research Ideas, PhD Project Plans, PostDoc Project Plans, Data Management Plans and Software Management Plans.

pricing-1

Very small outcomes have very small prices. We think producing a professional, machine-readable version of your work, assigning it a DOI, and enabling optional community peer review is something worth paying for. At € 50, single-figure publications (see Do and Mobley, 2015) are particularly appealingly priced, so as to encourage the community to explore this new format.  

pricing-2

For those who like to compare rates between different open-access journals, you’ll find our research article price very competitive at only €550. Unlike many open access journals, we accept review articles, too, as well as a wide variety of other research outputs.

pricing-3

Finally, you might be worried that charging for each and every single output could get quite expensive over time for an individual or an institution. We’ve got that covered with our package plans. For individuals, if you commit to buying a package upfront, we’re happy to reduce the overall price per output published to an average price between EUR 140 and 330 per article (see table below). For funders, institutions, conference organisers and others, we’re happy to make even more competitive deals; just contact us to discuss your needs at: rio@riojournal.com.

pricing-4

For further details, please see the Article Processing Charges and Promotions entries on the RIO website.

 

PART 2: The ARPHA Writing Tool: Adding Value to RIO

Last week, we presented some basic features of the ARPHA Writing Tool (AWT), focusing on the collaborative authoring process. This post will add the novelties that ARPHA has adopted when it comes to submission and post-submission processes.

arpha-validation

   

Two-step VALIDATION:

Checks are a must whenever high-quality, professional writing is concerned. Shortly before submitting their manuscripts via ARPHA, the authors are free to initiate the first step of the validation process at any time. As a result, an automated technical check runs through the manuscript to verify its consistency and compliance with the JATS (Journal Article Tag Suite) standard as well as key elements of the journal’s policy.

After the automated checks, there is a second-layer of human checks made by RIO editorial assistants in order to bring the manuscript to a pre-publication level. They check compliance with further elements of the journal’s policy as well as linguistic consistency and work with the authors before finalising it altogether.

 

After SUBMISSION:

Even after submission, collaboration continues. With the peer-review stage being as transparent and open as the whole vision of RIO is, this step is no less straightforward and organised. All peer reviews are automatically consolidated into a single online file that makes the editorial process simple and pleasant. This is made possible by the XML-based workflow in the AWT.

arpha-review-

PEER-REVIEW evolution:

Last week, we mentioned that an author can invite reviewers during the authoring stage itself. Let us elaborate here by saying that the ARPHA Writing Tool provides a functionality for pre-submission peer review(s) performed during the authoring process. These peer reviews are submitted together with the manuscript, so that the editorial evaluation and publication can be significantly sped up.

Post-publication UPDATES:

Even after an article is brought to life, ARPHA continues to play a role in it. Because of the XML workflow, authors are able to publish updated versions of their articles at any time. Once they request such an update, their work is returned to the ARPHA Writing Tool, where authors and peers are to collaborate once again. Eventually, the two versions are linked via CrossMark, so that nothing is lost.   

As you might have already noticed, just like RIO, the new ARPHA Writing Tool is constructed entirely around the needs of its authors. Why not find out yourself by giving ARPHA & RIO a try when they open for submission early next month?

 

PART 1: The ARPHA Writing Tool: Adding Value to RIO

The online, collaborative ARPHA Writing Tool (AWT) is an innovative, entirely Web-based, authoring, submission, peer-review, publication & dissemination authoring platform. While flexible to accommodate more traditional journal concepts, it clearly fits neatly with RIO’s novel approach to research publishing.

In the following two-part blog post, we will explain how the newly updated AWT will be integrated with RIO to provide a number of great features for unique user-experience.

ARPHA_part_1 (2)

All About User Friendliness:

The ARPHA Writing Tool did not appear out of nowhere. Pensoft has been developing its own writing tool since 2012, and the AWT is an updated version of this prototype, the Pensoft Writing Tool. The AWT builds onthe experience we have accumulated during the 3-year lifespan of its predecessor, and is designed to make the user experience as seamless and pleasant as possible.

How we did that? The AWT features a modern and intuitive design, where all features are listed and just a click away for the authors to use. The tool runs in one mode, showing real-time changes, comments and updates. Comments are left inline and automatically navigate through the text to make review and correcting easy.

awt-screen-shot (2)

A New Level of COLLABORATION:

RIO is all about collaboration, and within the ARPHA Writing Tool, this is now easier than ever. Authors can not only work collaboratively on a manuscript with their co-authors, but can also invite external contributors, such as mentors, pre-submission reviewers, linguistic and copy editors, or just colleagues, who may modify the manuscript or comment on it before submission. The system allows for such external contributors to not be listed as co-authors of the manuscript.

ARPHA_part_1 (3)

The collaborative peer-review process provides easy communication through track-change functionality, comments and replies, and automated, but customizable email and social network notification tools.

TEMPLATES for all needs

The AWT provides a large set of predefined, but flexible article templates covering many types of research outcomes. At the moment, the list fits the needs of the existing Pensoft journals, but with RIO coming soon, the choice of templates will grow to accommodate the needs of publishing full research cycles across a wide range of disciplines.

ARPHA_part_1 (4)

Adding CONTENT and REFERENCING made easy

A set of Web-based services and tools in AWT allows, e.g. for search and import of literature or data references, cross-referencing of in-text citations, import of tables, upload of images and multimedia, assembling images for display as reusably composite figures,  and even add videos to your article.

ARPHA_part_1

All the author needs to do to add content to an article is to select the right type from the side menu and then navigate through the instructions in the pop-up window. If in doubt, the authors can also consult the ‘Tips and Tricks’ on our website.

So far, we have explained some of the main features of the AWT that are useful throughout the authoring process, but there is much more to explore! In next week’s post, we will look at what happens after your manuscript is completed and ready for submission.